[go: up one dir, main page]

CN118799813A - Automatic identification method of airport support nodes based on relative position learning - Google Patents

Automatic identification method of airport support nodes based on relative position learning Download PDF

Info

Publication number
CN118799813A
CN118799813A CN202411261329.3A CN202411261329A CN118799813A CN 118799813 A CN118799813 A CN 118799813A CN 202411261329 A CN202411261329 A CN 202411261329A CN 118799813 A CN118799813 A CN 118799813A
Authority
CN
China
Prior art keywords
relative position
airport
vehicle
aircraft
features
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202411261329.3A
Other languages
Chinese (zh)
Other versions
CN118799813B (en
Inventor
党婉丽
罗谦
郑怀宇
耿龙
曹利波
王朝
牛杰
王江
裴翔宇
但军
刘晨
张启俊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Civil Aviation Electronic Technology Co ltd
Second Research Institute of CAAC
Original Assignee
Civil Aviation Electronic Technology Co ltd
Second Research Institute of CAAC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Civil Aviation Electronic Technology Co ltd, Second Research Institute of CAAC filed Critical Civil Aviation Electronic Technology Co ltd
Priority to CN202411261329.3A priority Critical patent/CN118799813B/en
Publication of CN118799813A publication Critical patent/CN118799813A/en
Application granted granted Critical
Publication of CN118799813B publication Critical patent/CN118799813B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/061Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using biological neurons, e.g. biological neurons connected to an integrated circuit
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/766Arrangements for image or video recognition or understanding using pattern recognition or machine learning using regression, e.g. by projecting features on hyperplanes
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Multimedia (AREA)
  • Biophysics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Databases & Information Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Neurology (AREA)
  • Microelectronics & Electronic Packaging (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开了基于相对位置学习的机场保障节点自动识别方法,涉及机坪保障技术领域,包括:S1、根据保障车辆和飞机之间的相对位置关系,制作标注保障车辆和飞机相对位置数据的数据集;S2、利用深度学习对数据集进行训练得到预测模型;S3、使用改进YOLOV7模型对机场实时传入的图像数据进行检测,得到飞机的关键点位置数据;S4、将关键点位置数据输入预测模型,完成机场保障节点自动识别,本发明弥补了完全依靠图像检测进行保障节点识别方法的不足,提高保障节点检测效率。

The present invention discloses an automatic identification method for airport support nodes based on relative position learning, which relates to the field of apron support technology, including: S1, according to the relative position relationship between the support vehicle and the aircraft, a data set is produced to mark the relative position data of the support vehicle and the aircraft; S2, the data set is trained by deep learning to obtain a prediction model; S3, the image data transmitted in real time from the airport is detected by an improved YOLOV7 model to obtain the key point position data of the aircraft; S4, the key point position data is input into the prediction model to complete the automatic identification of the airport support nodes. The present invention makes up for the shortcomings of the support node identification method that completely relies on image detection, and improves the detection efficiency of the support node.

Description

基于相对位置学习的机场保障节点自动识别方法Automatic identification method of airport support nodes based on relative position learning

技术领域Technical Field

本发明涉及机坪保障技术领域,特别涉及基于相对位置学习的机场保障节点自动识别方法。The invention relates to the technical field of apron support, and in particular to an automatic identification method of airport support nodes based on relative position learning.

背景技术Background Art

机场保障节点系统是现代机场运营中至关重要的一部分,它涵盖了一系列技术和设施,旨在确保机场各项运营活动的高效性、安全性和可靠性。在传统机场运行管理中,保障节点信息的采集通常依赖于人工观察和手动填报,这种方式存在一些问题。仅凭借人工观察和手动填报,会面临以下问题:增加了人员的工作负担,可能会影响工作质量;数据的实时性较差,可能导致信息不及时;手动填报容易出现漏报、误报等问题,使得数据的准确性和参考性受到影响。而机场保障节点系统就可以解决上述问题,通过多个子系统和设备的集成进行实时的检测和反馈协同,保障机场的秩序。The airport security node system is a vital part of modern airport operations. It covers a series of technologies and facilities to ensure the efficiency, safety and reliability of various airport operations. In traditional airport operation management, the collection of security node information usually relies on manual observation and manual reporting, which has some problems. Relying solely on manual observation and manual reporting will face the following problems: increasing the workload of personnel, which may affect the quality of work; the real-time performance of data is poor, which may lead to untimely information; manual reporting is prone to omissions, false reports and other problems, which affects the accuracy and reference of the data. The airport security node system can solve the above problems, and through the integration of multiple subsystems and equipment, real-time detection and feedback coordination can be carried out to ensure the order of the airport.

目前,图像处理技术在航班保障中起着十分重要的作用。通过智能感知技术,利用传感器网络和监控摄像头等设备,实时监测机场各个区域的情况。将收集到图像数据通过图像识别技术能够检测到飞机、车辆、行李和人员的位置和运动,以及检测环境因素如天气、温度和能见度等信息。建立可靠的通信和网络基础设施,保障各个设备之间的信息交换和互联。这包括局域网、无线网络、卫星通信等技术,确保系统在复杂的环境中稳定运行。与航空管制部门和航空公司的系统进行集成,确保航班安全和正常运行。At present, image processing technology plays a very important role in flight support. Through intelligent sensing technology, sensor networks and surveillance cameras are used to monitor the situation in various areas of the airport in real time. The collected image data can be used to detect the location and movement of aircraft, vehicles, luggage and personnel through image recognition technology, as well as environmental factors such as weather, temperature and visibility. Establish a reliable communication and network infrastructure to ensure information exchange and interconnection between various devices. This includes technologies such as local area networks, wireless networks, and satellite communications to ensure the stable operation of the system in complex environments. Integrate with the systems of air traffic control departments and airlines to ensure flight safety and normal operation.

现有基于图像识别的保障节点自动录入需要基于视频识别出保障车辆的类型,并且能够在整个视频过程中对车辆进行跟踪,但这种技术在保障车辆被遮挡或雨雾光照情况下可能出现严重误差,导致整个系统的识别效果大大降低。The existing automatic entry of support nodes based on image recognition requires identifying the type of support vehicle based on the video and being able to track the vehicle during the entire video process. However, this technology may have serious errors when the support vehicle is obscured or exposed to rain, fog and light, which greatly reduces the recognition effect of the entire system.

发明内容Summary of the invention

针对现有技术中的上述不足,本发明提供的相对位置学习的机场保障节点自动识别方法并不直接识别保障车辆,而是先识别出飞机的位姿,基于飞机位姿预测出保障车辆与飞机的相对位置,只要有车辆进入到识别区域,则可以认为是对应保障车辆的出现,由于本方法只需要基于图像识别技术识别出飞机上几个关键的部位,如机头和机尾,这两个关键部位的可辨识度很高,在雨雾或者夜间也能够很好地被识别出来,而且几乎不存在被遮挡的问题,其解决了现有的基于图像识别的保障节点自动录入方法在图像识别效果不佳的情况下无法进行保障节点检测的问题。In view of the above-mentioned deficiencies in the prior art, the airport support node automatic identification method based on relative position learning provided by the present invention does not directly identify the support vehicle, but first identifies the position and posture of the aircraft, and predicts the relative position of the support vehicle and the aircraft based on the position and posture of the aircraft. As long as a vehicle enters the identification area, it can be considered as the appearance of the corresponding support vehicle. Since the method only needs to identify several key parts of the aircraft based on image recognition technology, such as the nose and the tail, the recognizability of these two key parts is very high, and they can be well identified even in rain and fog or at night, and there is almost no problem of being blocked. It solves the problem that the existing support node automatic entry method based on image recognition cannot perform support node detection when the image recognition effect is poor.

为了达到上述发明目的,本发明采用的技术方案为:基于相对位置学习的机场保障节点自动识别方法,包括:In order to achieve the above-mentioned invention object, the technical solution adopted by the present invention is: an automatic identification method of airport support nodes based on relative position learning, comprising:

S1、根据保障车辆和飞机之间的相对位置关系,制作标注保障车辆和飞机相对位置数据的数据集;S1. Create a data set with the relative position data of the support vehicle and the aircraft according to the relative position relationship between the support vehicle and the aircraft;

S2、利用深度学习对数据集进行训练得到预测模型;S2. Use deep learning to train the data set to obtain a prediction model;

S3、使用改进YOLOV7模型对机场实时传入的图像数据进行检测,得到飞机的关键点位置数据;S3. Use the improved YOLOV7 model to detect the real-time image data transmitted from the airport to obtain the key point position data of the aircraft;

S4、将关键点位置数据输入预测模型,完成机场保障节点自动识别。S4. Input the key point location data into the prediction model to complete the automatic identification of airport support nodes.

进一步地:所述S1包括:Further: S1 includes:

S11、对飞机的3D图像进行不同拍摄角度以及拍摄距离的随机变换,并进行数据关键点标注,得到飞机位姿图像数据;S11, randomly transforming the 3D image of the aircraft at different shooting angles and shooting distances, and marking key points of the data to obtain aircraft posture image data;

S12、根据飞机位姿图像数据和对应的保障车辆位置,获取相对位置特征,并将相对位置特征作为标注保障车辆和飞机相对位置数据的数据集。S12. Obtain relative position features based on the aircraft posture image data and the corresponding support vehicle position, and use the relative position features as a data set for annotating the relative position data of the support vehicle and the aircraft.

进一步地:所述S12中,相对位置特征包括机头x坐标、机头y坐标、机尾x坐标、机尾y坐标、飞机朝向、保障车辆检测框左上角坐标和保障车辆检测框右下角坐标。Further: in S12, the relative position features include the nose x-coordinate, the nose y-coordinate, the tail x-coordinate, the tail y-coordinate, the aircraft orientation, the upper left corner coordinate of the support vehicle detection frame, and the lower right corner coordinate of the support vehicle detection frame.

进一步地:所述S2中,预测模型包括依次连接的输入层、第一隐藏层、第二隐藏层和输出层;Further: in S2, the prediction model includes an input layer, a first hidden layer, a second hidden layer and an output layer connected in sequence;

所述第一隐藏层包括64个神经元,所述第二隐藏层包括32个神经元。The first hidden layer includes 64 neurons, and the second hidden layer includes 32 neurons.

进一步地:所述S3中,改进YOLOV7模型对机场实时传入的图像数据的处理方法包括:Further: In S3, the method for improving the YOLOV7 model to process the image data transmitted in real time from the airport includes:

S31、将机场实时传入的图像数据输入改进YOLOV7模型,通过改进YOLOV7模型的主干网络和颈部网络,提取三种不同尺寸的特征图;S31. Input the real-time image data from the airport into the improved YOLOV7 model, and extract feature maps of three different sizes by improving the backbone network and neck network of the YOLOV7 model;

S32、使用解耦头分别对三种不同尺寸的特征图进行预测,得到用于分类的Cls特征、用于定位的Reg特征和用于置信度任务的Obj特征;S32. Use the decoupling head to predict the feature maps of three different sizes respectively, and obtain the Cls feature for classification, the Reg feature for positioning, and the Obj feature for the confidence task;

S33、将用于分类的Cls特征、用于定位的Reg特征和用于置信度任务的Obj特征经过堆叠操作组合为最终特征图,并将最终特征图作为飞机的关键点位置数据。S33, the Cls feature used for classification, the Reg feature used for positioning, and the Obj feature used for the confidence task are combined into a final feature map through a stacking operation, and the final feature map is used as the key point position data of the aircraft.

进一步地:所述S32中,解耦头包括分类分支和回归分支;Further: in said S32, the decoupling head includes a classification branch and a regression branch;

所述分类分支包括依次连接的两层3×3 卷积层和一层1×1卷积层;The classification branch includes two 3×3 convolutional layers and one 1×1 convolutional layer connected in sequence;

所述分类分支输出用于分类的Cls特征;The classification branch outputs a Cls feature for classification;

所述回归分支包括依次连接两层3×3 卷积层和平行的一层1×1卷积层;The regression branch includes two 3×3 convolutional layers connected in sequence and a 1×1 convolutional layer in parallel;

所述回归分支输出用于定位的Reg特征和用于置信度任务的Obj特征。The regression branch outputs Reg features for positioning and Obj features for confidence tasks.

进一步地:所述S4包括:Further: S4 includes:

S41、将关键点位置数据输入预测模型,通过预测模型输出保障车辆的检测框;S41, inputting the key point position data into the prediction model, and outputting the detection frame of the security vehicle through the prediction model;

S42、跟踪保障车辆的检测框中保障车辆段运动轨迹,根据保障车辆段运动轨迹,得到机场保障的开始时间节点和结束时间节点,完成机场保障节点自动识别。S42, tracking the movement trajectory of the support vehicle segment in the detection frame of the support vehicle, obtaining the start time node and the end time node of the airport support according to the movement trajectory of the support vehicle segment, and completing the automatic identification of the airport support node.

本发明的有益效果为:The beneficial effects of the present invention are:

本发明提供的相对位置学习的机场保障节点自动识别方法并不直接识别保障车辆,而是先识别出飞机的位姿,基于飞机位姿预测出保障车辆与飞机的相对位置,只要有车辆进入到识别区域,则可以认为是对应保障车辆的出现;The airport support node automatic identification method based on relative position learning provided by the present invention does not directly identify the support vehicle, but first identifies the posture of the aircraft, and predicts the relative position of the support vehicle and the aircraft based on the posture of the aircraft. As long as a vehicle enters the identification area, it can be considered that the corresponding support vehicle appears.

由于本方法只需要基于图像识别技术识别出飞机上几个关键的部位,如机头个机尾,这两个关键部位的可辨识度很高,在雨雾或者夜间也能够很好地被识别出来,而且几乎不存在被遮挡的问题,其解决了现有的基于图像识别的保障节点自动录入方法在图像识别效果不佳的情况下无法进行保障节点检测的问题。Since this method only needs to identify several key parts of the aircraft based on image recognition technology, such as the nose and the tail, these two key parts have high identifiability and can be well identified in rain, fog or at night, and there is almost no problem of being blocked. It solves the problem that the existing automatic entry method of support nodes based on image recognition cannot perform support node detection when the image recognition effect is poor.

附图说明BRIEF DESCRIPTION OF THE DRAWINGS

图1为基于相对位置学习的机场保障节点自动识别方法流程图。FIG1 is a flow chart of an automatic identification method for airport support nodes based on relative position learning.

图2为不同远近的飞机3D模型关键点标注示意图。Figure 2 is a schematic diagram of key point annotation of the 3D model of an aircraft at different distances.

图3为不同俯视角度的飞机3D模型关键点标注示意图。FIG3 is a schematic diagram of key point annotation of the aircraft 3D model at different top-down angles.

图4为飞机相对位置的正反面标注示意图。Figure 4 is a schematic diagram of the front and back sides of the aircraft relative positions.

图5为预测模型结构示意图。FIG5 is a schematic diagram of the prediction model structure.

图6为基于相对位置学习的机场保障节点自动识别方法原理图。FIG6 is a schematic diagram showing the principle of an automatic identification method for airport support nodes based on relative position learning.

具体实施方式DETAILED DESCRIPTION

下面对本发明的具体实施方式进行描述,以便于本技术领域的技术人员理解本发明,但应该清楚,本发明不限于具体实施方式的范围,对本技术领域的普通技术人员来讲,只要各种变化在所附的权利要求限定和确定的本发明的精神和范围内,这些变化是显而易见的,一切利用本发明构思的发明创造均在保护之列。The specific implementation modes of the present invention are described below so that those skilled in the art can understand the present invention. However, it should be clear that the present invention is not limited to the scope of the specific implementation modes. For those of ordinary skill in the art, as long as various changes are within the spirit and scope of the present invention as defined and determined by the attached claims, these changes are obvious, and all inventions and creations utilizing the concept of the present invention are protected.

如图1所示,在本发明的一个实施例中,提供基于相对位置学习的机场保障节点自动识别方法,包括:As shown in FIG1 , in one embodiment of the present invention, a method for automatically identifying airport support nodes based on relative position learning is provided, comprising:

S1、根据保障车辆和飞机之间的相对位置关系,制作标注保障车辆和飞机相对位置数据的数据集;S1. Create a data set with the relative position data of the support vehicle and the aircraft according to the relative position relationship between the support vehicle and the aircraft;

S2、利用深度学习对数据集进行训练得到预测模型;S2. Use deep learning to train the data set to obtain a prediction model;

S3、使用改进YOLOV7模型对机场实时传入的图像数据进行检测,得到飞机的关键点位置数据;S3. Use the improved YOLOV7 model to detect the real-time image data transmitted from the airport to obtain the key point position data of the aircraft;

S4、将关键点位置数据输入预测模型,完成机场保障节点自动识别。S4. Input the key point location data into the prediction model to complete the automatic identification of airport support nodes.

所述S1包括:The S1 includes:

S11、对飞机的3D图像进行不同拍摄角度以及拍摄距离的随机变换,并进行数据关键点标注,得到飞机位姿图像数据;S11, randomly transforming the 3D image of the aircraft at different shooting angles and shooting distances, and marking key points of the data to obtain aircraft posture image data;

S12、根据飞机位姿图像数据和对应的保障车辆位置,获取相对位置特征,并将相对位置特征作为标注保障车辆和飞机相对位置数据的数据集。S12. Obtain relative position features based on the aircraft posture image data and the corresponding support vehicle position, and use the relative position features as a data set for annotating the support vehicle and aircraft relative position data.

在本发明的一个实施例中,如图2和图3所示,所述S11中,图2展示了不同远近的飞机3D模型关键点标注示意图,图2(a)代表拍摄距离为近距离、图2(b)代表拍摄距离为中等距离、图2(c)代表拍摄距离为远距离;In one embodiment of the present invention, as shown in FIG. 2 and FIG. 3 , in the S11, FIG. 2 shows a schematic diagram of key point annotation of a 3D model of an aircraft at different distances, FIG. 2 (a) represents a shooting distance of a short distance, FIG. 2 (b) represents a shooting distance of a medium distance, and FIG. 2 (c) represents a shooting distance of a long distance;

图3展示了在俯视角分别为0°、30°、60°三个角度的3D模型截图,图3(a)代表俯视角度为0°,图3(b)代表俯视角度为30°,图3(c)代表俯视角度为60°;Figure 3 shows screenshots of the 3D model at three top-view angles of 0°, 30°, and 60°. Figure 3(a) represents a top-view angle of 0°, Figure 3(b) represents a top-view angle of 30°, and Figure 3(c) represents a top-view angle of 60°.

对每一个角度与每一个距离的不同组合都生成相应的标注数据,标注出飞机机头、飞机机尾的位置以及不同保障车辆在当前飞机位姿情况下的相对位置,将这些标注出的位置数据作为保障车辆位置预测的训练数据。Corresponding annotation data is generated for each different combination of angle and distance, marking the position of the aircraft nose and tail as well as the relative positions of different support vehicles in the current aircraft posture. These annotated position data are used as training data for support vehicle position prediction.

在本发明的一个实施例中,所述S12中,相对位置特征选择的过程是一项精细而关键的任务,包含飞机和保障车辆相对位置的筛选,通常涉及机头位置、机尾位置、客梯车位置等不同类型的相对位置特征。在进行特征选择时,首先需要对图像或视频进行观察和分析,以确定每个特征的位置和类型。接着,根据目标的形状、大小和位置等特征,采用不同的计算公式来确定其具体的标注位置和属性。In one embodiment of the present invention, the relative position feature selection process in S12 is a delicate and critical task, which includes the screening of the relative positions of the aircraft and the support vehicles, and usually involves different types of relative position features such as the nose position, the tail position, and the passenger elevator position. When performing feature selection, it is first necessary to observe and analyze the image or video to determine the position and type of each feature. Then, according to the shape, size, position and other characteristics of the target, different calculation formulas are used to determine its specific annotation position and attributes.

在本实施例中,选取的相对位置特征包括机头x坐标、机头y坐标、机尾x坐标、机尾y坐标、飞机朝向、保障车辆检测框左上角坐标和保障车辆检测框右下角坐标;In this embodiment, the selected relative position features include the nose x-coordinate, nose y-coordinate, tail x-coordinate, tail y-coordinate, aircraft orientation, upper left corner coordinate of the support vehicle detection frame, and lower right corner coordinate of the support vehicle detection frame;

由于图像的原点在图像左上角,我们以机头标注框中心点到图片左上角x坐标的距离作为机头x坐标,以机头标注框中心点到图片左上角y坐标的距离作为机头y坐标,以机尾标注框中心点到图片左上角x坐标的距离作为机尾x坐标,以机尾标注框中心点到图片左上角y坐标的距离作为机尾y坐标;Since the origin of the image is in the upper left corner of the image, we use the distance from the center point of the nose annotation box to the x coordinate of the upper left corner of the image as the x coordinate of the nose, the distance from the center point of the nose annotation box to the y coordinate of the upper left corner of the image as the y coordinate of the nose, the distance from the center point of the tail annotation box to the x coordinate of the upper left corner of the image as the x coordinate of the tail, and the distance from the center point of the tail annotation box to the y coordinate of the upper left corner of the image as the y coordinate of the tail;

所以机头的x坐标为机头标注框中心点的x坐标;机头的y坐标为机头标注框中心点的y坐标;机尾的x坐标为机尾标注框中心点的x坐标,机尾的y坐标为机尾标注框中心点的y坐标;Therefore, the x-coordinate of the nose is the x-coordinate of the center point of the nose annotation box; the y-coordinate of the nose is the y-coordinate of the center point of the nose annotation box; the x-coordinate of the tail is the x-coordinate of the center point of the tail annotation box, and the y-coordinate of the tail is the y-coordinate of the center point of the tail annotation box;

如图4所示,飞机朝向的特征选取中,图4(a),将机头朝向视频的机头,标注为正向机头,机头朝视频的机尾标注为正向机尾,图4(b),将机尾朝向视频的机头标注为反向机头,机尾朝视频的机尾标注为反向机尾,增强客梯车的识别准确率。As shown in Figure 4, in the feature selection of the aircraft orientation, Figure 4 (a) marks the nose of the aircraft facing the video as the positive nose, and the nose of the aircraft facing the video as the positive tail. Figure 4 (b) marks the nose of the aircraft facing the video as the reverse nose, and the tail of the aircraft facing the video as the reverse tail, so as to enhance the recognition accuracy of the passenger elevator.

由于需要预测保障车辆与飞机的相对位置为保障车辆的检测框位置,而检测框只需要确定检测框左上角坐标和检测框右下角坐标,值得指出的是我们不容易预测保障车辆在图片中的绝对位置,但较容易预测保障车辆与飞机的相对位置,所以检测框左上角坐标和检测框右下角坐标是检测框与飞机机头的相对位置;Since the relative position of the support vehicle and the aircraft needs to be predicted as the detection frame position of the support vehicle, and the detection frame only needs to determine the coordinates of the upper left corner and the lower right corner of the detection frame, it is worth pointing out that it is not easy to predict the absolute position of the support vehicle in the picture, but it is easier to predict the relative position of the support vehicle and the aircraft. Therefore, the coordinates of the upper left corner and the lower right corner of the detection frame are the relative positions of the detection frame and the nose of the aircraft;

保障车辆的检测是通过确定保障车辆的x坐标与y坐标,检测框是通过保障车辆x,y坐标与宽度高度决定的,保障车辆的中心位置为:X保障车辆=X保障车辆x坐标-X机头The detection of the support vehicle is done by determining the x-coordinate and y-coordinate of the support vehicle. The detection frame is determined by the x-, y-coordinate and width and height of the support vehicle. The center position of the support vehicle is: X support vehicle = X support vehicle x-coordinate - X head ;

其中X保障车辆是保障车辆相对机头的x坐标,X保障车辆x坐标是保障车辆的绝对x坐标,X机头是机头的实际x坐标;Where Xsupportvehicle is the x-coordinate of the support vehicle relative to the nose, Xsupportvehiclexcoordinate is the absolute x-coordinate of the support vehicle, and Xnose is the actual x-coordinate of the nose;

保障车辆检测框左上角坐标=(X保障车辆- W保障车辆/2, y - H保障车辆/2),保障车辆检测框右下角坐标=(X保障车辆+ W保障车辆/2,y+ H保障车辆/2),其中W保障车辆是保障车辆检测框的宽度,H保障车辆是保障车辆检测框的高度,这样就可以得到左上角和右下角的坐标,从而确定整个保障车辆标检测框的位置。The coordinates of the upper left corner of the security vehicle detection frame = (X security vehicle - W security vehicle / 2, y - H security vehicle / 2), and the coordinates of the lower right corner of the security vehicle detection frame = (X security vehicle + W security vehicle / 2, y + H security vehicle / 2), where W security vehicle is the width of the security vehicle detection frame and H security vehicle is the height of the security vehicle detection frame. In this way, the coordinates of the upper left corner and the lower right corner can be obtained, thereby determining the position of the entire security vehicle detection frame.

通过图4(a)和图4(b)的标注,我们获得的相对位置训练数据如表1所示:Through the annotations of Figure 4(a) and Figure 4(b), we obtained the relative position training data as shown in Table 1:

表1基于3D模型的飞机及其保障车辆相对位置数据Table 1 Relative position data of aircraft and its support vehicles based on 3D model

序号Serial number 机头x坐标Head x coordinate 机头y坐标Head y coordinate 机尾x坐标Tail x coordinate 机尾y坐标Tail y coordinate 飞机朝向Aircraft direction 保障车辆检测框左上角坐标Ensure the coordinates of the upper left corner of the vehicle detection frame 保障车辆检测框右下角坐标Ensure the coordinates of the lower right corner of the vehicle detection frame 保障车辆类型Protected vehicle type 图4(a)Figure 4(a) 0.4932290.493229 0.4929690.492969 0.4973960.497396 0.5382810.538281 正向Positive 0.6143230.614323 0.5617190.561719 客梯车Passenger elevator 图4(b)Figure 4(b) 0.2710940.271094 0.5089840.508984 0.5997400.599740 0.5136720.513672 反向Reverse 0.3802080.380208 0.5648440.564844 客梯车Passenger elevator

如图5所示,所述S2中,预测模型包括依次连接的输入层、第一隐藏层、第二隐藏层和输出层;所述第一隐藏层包括64个神经元,所述第二隐藏层包括32个神经元。As shown in FIG. 5 , in S2 , the prediction model includes an input layer, a first hidden layer, a second hidden layer and an output layer connected in sequence; the first hidden layer includes 64 neurons, and the second hidden layer includes 32 neurons.

使用准备好的特征标注数据集,将预处理后的数据输入到预测模型中进行训练,通过反向传播算法更新权重,以最小化损失函数。模型通过反向传播算法更新权重,以最小化损失函数。Using the prepared feature annotation dataset, the preprocessed data is input into the prediction model for training, and the weights are updated through the back propagation algorithm to minimize the loss function. The model updates the weights through the back propagation algorithm to minimize the loss function.

如图6所示,上半部分描述了如何基于YOLOV7模型识别飞机的关键点数据,在输入端,批处理的图像数据经过特征处理后,并通过一系列数据增强技巧进行处理。主干网络(Backbone)是YOLOv7 模型的核心部分,负责对输入图像进行特征提取和表示。它通常由一系列的卷积层、池化层和其他基本操作组成,将输入图像转换为具有更高层次的语义信息的特征图。主干网络(Backbone)的主要作用是从原始图像中提取特征,使得模型能够理解图像中的内容和结构。As shown in Figure 6, the upper part describes how to identify the key point data of an aircraft based on the YOLOV7 model. At the input end, the batched image data is processed after feature processing and a series of data enhancement techniques. The backbone network is the core part of the YOLOv7 model, responsible for feature extraction and representation of the input image. It usually consists of a series of convolutional layers, pooling layers, and other basic operations to convert the input image into a feature map with higher-level semantic information. The main function of the backbone network is to extract features from the original image so that the model can understand the content and structure in the image.

颈部网络(Neck)位于主干网络后面,负责进一步处理和融合来自不同层次特征图的信息,以提高模型对目标的检测性能。颈部网络(Neck)的设计使用了特征融合、尺度变换、注意力机制等操作,旨在增强模型的感知能力和对目标的建模能力。The neck network is located behind the backbone network and is responsible for further processing and fusing information from feature maps at different levels to improve the model's detection performance of the target. The design of the neck network uses operations such as feature fusion, scale transformation, and attention mechanism to enhance the model's perception and target modeling capabilities.

由于目标检测领域包括分类和预测两项任务,分类任务主要关注特征层所提取的特征与真实类别特征的相似度,预测任务主要关注预测框与真实框的位置坐标从而进行预测边界框的参数修正。由于目标是采集到表1的数据,为了使表1的数据尽量准确,就需要使得预测边界框的参数修正尽量准确。而YOLO系列算法一直使用同一张特征图进行分类和定位任务,不同任务使用同样的共享参数,使两个任务无法专注于各自的目标,其检测效果会因此变差。为了分离分类和预测任务,本发明用解耦头(Decoupled Head)结构来替换YOLOv7中网络输出端的预测头。Since the field of target detection includes two tasks, classification and prediction, the classification task mainly focuses on the similarity between the features extracted by the feature layer and the real category features, and the prediction task mainly focuses on the position coordinates of the prediction box and the real box to correct the parameters of the prediction bounding box. Since the goal is to collect the data in Table 1, in order to make the data in Table 1 as accurate as possible, it is necessary to make the parameter correction of the prediction bounding box as accurate as possible. The YOLO series of algorithms have always used the same feature map for classification and positioning tasks, and different tasks use the same shared parameters, so that the two tasks cannot focus on their respective targets, and their detection effect will deteriorate. In order to separate the classification and prediction tasks, the present invention replaces the prediction head at the network output end in YOLOv7 with a decoupled head structure.

所述S3中,改进YOLOV7模型对机场实时传入的图像数据的处理方法包括:In S3, the method for improving the YOLOV7 model to process the image data transmitted in real time from the airport includes:

S31、将机场实时传入的图像数据输入改进YOLOV7模型,通过改进YOLOV7模型的主干网络和颈部网络,提取三种不同尺寸的特征图;S31. Input the real-time image data from the airport into the improved YOLOV7 model, and extract feature maps of three different sizes by improving the backbone network and neck network of the YOLOV7 model;

S32、使用解耦头分别对三种不同尺寸的特征图进行预测,得到用于分类的Cls特征、用于定位的Reg特征和用于置信度任务的Obj特征;S32. Use the decoupling head to predict the feature maps of three different sizes respectively, and obtain the Cls feature for classification, the Reg feature for positioning, and the Obj feature for the confidence task;

S33、将用于分类的Cls特征、用于定位的Reg特征和用于置信度任务的Obj特征经过堆叠操作组合为最终特征图,并将最终特征图作为飞机的关键点位置数据。S33, the Cls feature used for classification, the Reg feature used for positioning, and the Obj feature used for the confidence task are combined into a final feature map through a stacking operation, and the final feature map is used as the key point position data of the aircraft.

所述S32中,解耦头包括分类分支和回归分支;In the S32, the decoupling head includes a classification branch and a regression branch;

所述分类分支包括依次连接的两层3×3 卷积层和一层1×1卷积层;The classification branch includes two 3×3 convolutional layers and one 1×1 convolutional layer connected in sequence;

所述分类分支输出用于分类的Cls特征;The classification branch outputs a Cls feature for classification;

所述回归分支包括依次连接两层3×3 卷积层和平行的一层1×1卷积层;The regression branch includes two 3×3 convolutional layers connected in sequence and a 1×1 convolutional layer in parallel;

所述回归分支输出用于定位的Reg特征和用于置信度任务的Obj特征。The regression branch outputs Reg features for positioning and Obj features for confidence tasks.

所述S4包括:The S4 includes:

S41、将关键点位置数据输入预测模型,通过预测模型输出保障车辆的检测框;S41, inputting the key point position data into the prediction model, and outputting the detection frame of the security vehicle through the prediction model;

S42、跟踪保障车辆的检测框中保障车辆段运动轨迹,根据保障车辆段运动轨迹,得到机场保障的开始时间节点和结束时间节点,完成机场保障节点自动识别。S42, tracking the movement trajectory of the support vehicle segment in the detection frame of the support vehicle, obtaining the start time node and the end time node of the airport support according to the movement trajectory of the support vehicle segment, and completing the automatic identification of the airport support node.

一旦在预测出的相对位置检测到车辆,说明对应的保障车辆已经进入工作区域,就可以开始跟踪它们的运动轨迹,可以通过在连续帧之间匹配车辆的边界框来实现。当检测到车辆时,记录其出现的时间戳。同时,当车辆从图像或视频中消失时,也记录下相应的时间戳。分析这些记录的时间戳数据,可以得出车辆在预测位置出现和消失的时间。将检测到的各个车辆的出现时间,即各种车辆到位之后即可得到保障节点的开始时间,按照正常程序执行下去。在保障节点开始之后,各个车辆进行有序的入场,在登机完成之后,各个车辆进行有序撤离,得到各个车辆的消失时间之后,飞机起飞即可得到保障节点的结束时间。Once a vehicle is detected at the predicted relative position, it means that the corresponding support vehicle has entered the work area, and its movement trajectory can be tracked. This can be achieved by matching the bounding box of the vehicle between consecutive frames. When a vehicle is detected, the timestamp of its appearance is recorded. At the same time, when the vehicle disappears from the image or video, the corresponding timestamp is also recorded. By analyzing these recorded timestamp data, the time when the vehicle appears and disappears at the predicted position can be obtained. The appearance time of each detected vehicle, that is, after various vehicles are in place, the start time of the support node can be obtained, and it is executed according to the normal procedure. After the start of the support node, each vehicle enters in an orderly manner. After boarding is completed, each vehicle evacuates in an orderly manner. After the disappearance time of each vehicle is obtained, the end time of the support node can be obtained when the aircraft takes off.

以上实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围。The above embodiments are only used to illustrate the technical solutions of the present application, rather than to limit them. Although the present application has been described in detail with reference to the aforementioned embodiments, those skilled in the art should understand that they can still modify the technical solutions described in the aforementioned embodiments, or make equivalent replacements for some of the technical features therein. However, these modifications or replacements do not deviate the essence of the corresponding technical solutions from the spirit and scope of the technical solutions of the embodiments of the present application.

Claims (7)

1. The airport guarantee node automatic identification method based on relative position learning is characterized by comprising the following steps:
S1, manufacturing a data set for marking the relative position data of a guarantee vehicle and an airplane according to the relative position relation between the guarantee vehicle and the airplane;
s2, training the data set by deep learning to obtain a prediction model;
s3, detecting image data transmitted in real time to an airport by using an improved YOLOV model to obtain key point position data of the airplane;
and S4, inputting the position data of the key points into a prediction model to finish automatic identification of the airport guarantee nodes.
2. The method for automatically identifying airport security nodes based on relative position learning of claim 1, wherein S1 comprises:
s11, carrying out random transformation on the 3D images of the aircraft at different shooting angles and shooting distances, and carrying out data key point labeling to obtain pose image data of the aircraft;
s12, acquiring relative position features according to the pose image data of the aircraft and the corresponding position of the guarantee vehicle, and taking the relative position features as a data set for labeling the relative position data of the guarantee vehicle and the aircraft.
3. The method for automatically identifying airport security nodes based on relative position learning of claim 1, wherein in S12, the relative position features comprise a nose x coordinate, a nose y coordinate, a tail x coordinate, a tail y coordinate, an airplane orientation, an upper left corner coordinate of a security vehicle detection frame, and a lower right corner coordinate of a security vehicle detection frame.
4. The method for automatically identifying airport security nodes based on relative position learning according to claim 1, wherein in S2, the prediction model comprises an input layer, a first hidden layer, a second hidden layer and an output layer which are sequentially connected;
The first hidden layer includes 64 neurons and the second hidden layer includes 32 neurons.
5. The method for automatically identifying airport security nodes based on relative position learning according to claim 1, wherein in S3, the method for processing image data of an airport incoming in real time by using the improved YOLOV model comprises the following steps:
S31, inputting image data transmitted in real time to an airport into an improved YOLOV model, and extracting three feature images with different sizes by improving a main network and a neck network of the YOLOV model;
s32, respectively predicting three feature graphs with different sizes by using a decoupling head to obtain Cls features for classification, reg features for positioning and Obj features for confidence tasks;
S33, combining the Cls features for classification, the Reg features for positioning and the Obj features for confidence tasks into a final feature map through stacking operation, and taking the final feature map as the key point position data of the airplane.
6. The method for automatically identifying airport security nodes based on relative position learning of claim 5, wherein in S32, the decoupling head comprises a classification branch and a regression branch;
the classifying branch comprises two layers of 3 multiplied by 3 convolution layers and one layer of 1 multiplied by 1 convolution layer which are sequentially connected;
The classification branch outputs Cls characteristics for classification;
the regression branch comprises two layers of 3 multiplied by 3 convolution layers and a parallel layer of 1 multiplied by 1 convolution layer which are connected in sequence;
the regression branches output Reg features for localization and Obj features for confidence tasks.
7. The method for automatically identifying airport security nodes based on relative position learning of claim 1, wherein S4 comprises:
s41, inputting the position data of the key points into a prediction model, and outputting a detection frame for guaranteeing the vehicle through the prediction model;
S42, tracking a motion track of a guarantee vehicle section in a detection frame of the guarantee vehicle, obtaining a start time node and an end time node of airport guarantee according to the motion track of the guarantee vehicle section, and completing automatic identification of the airport guarantee node.
CN202411261329.3A 2024-09-10 2024-09-10 Automatic identification method of airport support nodes based on relative position learning Active CN118799813B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202411261329.3A CN118799813B (en) 2024-09-10 2024-09-10 Automatic identification method of airport support nodes based on relative position learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202411261329.3A CN118799813B (en) 2024-09-10 2024-09-10 Automatic identification method of airport support nodes based on relative position learning

Publications (2)

Publication Number Publication Date
CN118799813A true CN118799813A (en) 2024-10-18
CN118799813B CN118799813B (en) 2024-12-03

Family

ID=93031250

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202411261329.3A Active CN118799813B (en) 2024-09-10 2024-09-10 Automatic identification method of airport support nodes based on relative position learning

Country Status (1)

Country Link
CN (1) CN118799813B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN120472197A (en) * 2025-07-15 2025-08-12 中国民用航空总局第二研究所 Flight guarantee node relative position generation method based on 3D reconstruction

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6545601B1 (en) * 1999-02-25 2003-04-08 David A. Monroe Ground based security surveillance system for aircraft and other commercial vehicles
CN106251710A (en) * 2016-10-08 2016-12-21 合肥飞友网络科技有限公司 A kind of method calculating flight guarantee node time by vehicle location
CN115797822A (en) * 2022-10-10 2023-03-14 南京航空航天大学 A method and system for automatic identification of key time nodes in an airport scene
CN117151680A (en) * 2023-06-13 2023-12-01 大连理工大学 Method and system for inspecting bridge diseases based on YOLOv6
WO2024152477A1 (en) * 2023-01-17 2024-07-25 南京莱斯电子设备有限公司 Airport flight zone real-time target detection method based on multiscale feature decoupling

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6545601B1 (en) * 1999-02-25 2003-04-08 David A. Monroe Ground based security surveillance system for aircraft and other commercial vehicles
CN106251710A (en) * 2016-10-08 2016-12-21 合肥飞友网络科技有限公司 A kind of method calculating flight guarantee node time by vehicle location
CN115797822A (en) * 2022-10-10 2023-03-14 南京航空航天大学 A method and system for automatic identification of key time nodes in an airport scene
WO2024152477A1 (en) * 2023-01-17 2024-07-25 南京莱斯电子设备有限公司 Airport flight zone real-time target detection method based on multiscale feature decoupling
CN117151680A (en) * 2023-06-13 2023-12-01 大连理工大学 Method and system for inspecting bridge diseases based on YOLOv6

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
JICHAO ZHANG 等: "Optimization of Apron Support Vehicle Operation Scheduling Based on Multi-Layer Coding Genetic Algorithm", 《APPLIED SCIENCES》, 23 May 2022 (2022-05-23), pages 1 - 24 *
沈铖武 等: "挂弹车实时对准测量方法", 《测试技术学报》, vol. 29, no. 5, 30 October 2015 (2015-10-30), pages 427 - 431 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN120472197A (en) * 2025-07-15 2025-08-12 中国民用航空总局第二研究所 Flight guarantee node relative position generation method based on 3D reconstruction

Also Published As

Publication number Publication date
CN118799813B (en) 2024-12-03

Similar Documents

Publication Publication Date Title
EP3633615A1 (en) Deep learning network and average drift-based automatic vessel tracking method and system
CN108830171B (en) Intelligent logistics warehouse guide line visual detection method based on deep learning
CN113516664A (en) A Visual SLAM Method Based on Semantic Segmentation of Dynamic Points
CN103162682B (en) Based on the indoor path navigation method of mixed reality
CN112750150A (en) Vehicle flow statistical method based on vehicle detection and multi-target tracking
CN114116933B (en) A semantic-topological joint mapping method based on monocular images
CN113724293A (en) Vision-based intelligent internet public transport scene target tracking method and system
CN112069997B (en) A DenseHR-Net-based UAV autonomous landing target extraction method and device
CN111881749A (en) Bidirectional pedestrian flow statistical method based on RGB-D multi-modal data
CN112488061A (en) Multi-aircraft detection and tracking method combined with ADS-B information
CN112613668A (en) Scenic spot dangerous area management and control method based on artificial intelligence
WO2024216523A1 (en) Method and system for sensing foreign matter within urban rail train travellng clearance, and apparatus and medium
CN118799813A (en) Automatic identification method of airport support nodes based on relative position learning
CN114812398A (en) A high-precision real-time crack detection platform based on UAV
CN118675106A (en) Real-time monitoring method, system, device and storage medium for falling rocks based on machine vision
CN116664851A (en) Automatic driving data extraction method based on artificial intelligence
CN115018883A (en) An infrared autonomous inspection method for transmission line UAV based on optical flow and Kalman filtering
Wu et al. Multi camera localization handover based on YOLO object detection algorithm in complex environments
Liu et al. A lightweight LiDAR-camera sensing method of obstacles detection and classification for autonomous rail rapid transit
CN118155148A (en) A method and system for intelligently detecting emergencies in a running tunnel
CN120451046A (en) A bridge crack detection method and system combining drones and computer vision
Zhang et al. Accurate Detection and Tracking of Small‐Scale Vehicles in High‐Altitude Unmanned Aerial Vehicle Bird‐View Imagery
CN119691513A (en) Dynamic risk assessment method combining visual positioning and mapping with control barrier function
CN117746524B (en) Security inspection system and method based on SLAM and crowd abnormal behavior identification
CN118172385B (en) Unmanned aerial vehicle aerial video track data extraction method for traffic participants at complex intersections

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant