CN117901892A - Automatic driving control method, device, electronic device and storage medium - Google Patents
Automatic driving control method, device, electronic device and storage medium Download PDFInfo
- Publication number
- CN117901892A CN117901892A CN202311568377.2A CN202311568377A CN117901892A CN 117901892 A CN117901892 A CN 117901892A CN 202311568377 A CN202311568377 A CN 202311568377A CN 117901892 A CN117901892 A CN 117901892A
- Authority
- CN
- China
- Prior art keywords
- features
- driving control
- training sample
- parameters
- cognitive
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W60/00—Drive control systems specially adapted for autonomous road vehicles
- B60W60/001—Planning or execution of driving tasks
Landscapes
- Engineering & Computer Science (AREA)
- Automation & Control Theory (AREA)
- Human Computer Interaction (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Traffic Control Systems (AREA)
Abstract
Description
技术领域Technical Field
本申请涉及车辆技术领域,尤其涉及一种自动驾驶控制方法、装置、电子设备及存储介质。The present application relates to the field of vehicle technology, and in particular to an automatic driving control method, device, electronic device and storage medium.
背景技术Background Art
随着科技的不断发展,算力的持续突破,自动驾驶迅猛发展,自动驾驶可以为人们提供更安全、便捷和高效的出行方式,同时也对整个交通系统和城市规划带来了深远的影响。With the continuous development of science and technology, continuous breakthroughs in computing power, and the rapid development of autonomous driving, autonomous driving can provide people with a safer, more convenient and efficient way of travel, while also having a profound impact on the entire transportation system and urban planning.
自动驾驶系统主流的技术方案主要分为两种:基于规则分模块架构和基于数据驱动的端到端感知决策架构。虽然基于规则分模块的架构凭借算力需求低、部署简单、过程可解释等优势已经实现量产落地,但是难以实现全场景域覆盖,分模块求解的结果并不是全局最优解。端到端方案以原始传感器数据为输入,并生成规划和/或低级控制动作作为输出,可以进行联合的、全局的优化,自动驾驶决策的准确性较高。The mainstream technical solutions for autonomous driving systems are mainly divided into two types: rule-based modular architecture and data-driven end-to-end perception decision architecture. Although the rule-based modular architecture has been put into mass production due to its advantages such as low computing power requirements, simple deployment, and explainable processes, it is difficult to achieve full scene coverage, and the results of modular solutions are not the global optimal solution. The end-to-end solution uses raw sensor data as input and generates planning and/or low-level control actions as output. It can perform joint and global optimization, and the accuracy of autonomous driving decisions is relatively high.
然而,端到端感知决策架构的可解释性较差,因此在控制车辆执行突发性行为时,容易导致乘坐自动驾驶车辆的用户无法理解自动驾驶车辆的驾驶行为而产生疑惑、恐慌、不安等负面情绪,用户体验感较差。However, the end-to-end perception decision-making architecture has poor interpretability. Therefore, when controlling the vehicle to perform sudden behaviors, it is easy for users riding in the autonomous driving vehicle to not understand the driving behavior of the autonomous driving vehicle and experience negative emotions such as confusion, panic, and anxiety, resulting in a poor user experience.
发明内容Summary of the invention
本申请的主要目的在于提供一种自动驾驶控制方法、装置、电子设备及存储介质,旨在解决相关技术中端到端感知决策方案的可解释性较差的技术问题。The main purpose of this application is to provide an autonomous driving control method, device, electronic device and storage medium, aiming to solve the technical problem of poor interpretability of end-to-end perception decision-making solutions in related technologies.
为实现上述目的,本申请提供一种自动驾驶控制方法,所述自动驾驶控制方法采用认知决策模型,所述认知决策模型包括认知层和决策层;所述自动驾驶控制方法包括以下步骤:To achieve the above-mentioned purpose, the present application provides an automatic driving control method, wherein the automatic driving control method adopts a cognitive decision model, wherein the cognitive decision model includes a cognitive layer and a decision layer; the automatic driving control method includes the following steps:
获取车内外监测数据,通过所述认知层从所述车内外监测数据中提取当前认知层特征;Acquire vehicle interior and exterior monitoring data, and extract current cognitive layer features from the vehicle interior and exterior monitoring data through the cognitive layer;
将所述当前认知层特征输入所述决策层,确定自动驾驶控制参数;Inputting the current cognitive layer features into the decision layer to determine the automatic driving control parameters;
根据所述当前认知层特征和所述自动驾驶控制参数,生成驾驶行为解释。Generate a driving behavior explanation based on the current cognitive layer characteristics and the autonomous driving control parameters.
本申请还提供一种自动驾驶控制装置,所述自动驾驶控制装置上部署有认知决策模型,所述认知决策模型包括认知层和决策层,所述自动驾驶控制装置包括:The present application also provides an automatic driving control device, on which a cognitive decision model is deployed, the cognitive decision model includes a cognitive layer and a decision layer, and the automatic driving control device includes:
认知模块,用于获取车内外监测数据,通过所述认知层从所述车内外监测数据中提取当前认知层特征;A cognitive module, used for acquiring in-vehicle and out-vehicle monitoring data, and extracting current cognitive layer features from the in-vehicle and out-vehicle monitoring data through the cognitive layer;
决策模块,用于将所述当前认知层特征输入所述决策层,确定自动驾驶控制参数;A decision module, used for inputting the current cognitive layer features into the decision layer to determine the automatic driving control parameters;
解释模块,用于根据所述当前认知层特征和所述自动驾驶控制参数,生成驾驶行为解释。An explanation module is used to generate a driving behavior explanation based on the current cognitive layer characteristics and the automatic driving control parameters.
本申请还提供一种电子设备,所述电子设备为实体设备,所述电子设备包括:存储器、处理器以及存储在所述存储器上并可在所述处理器上运行的所述的自动驾驶控制方法的程序,所述的自动驾驶控制方法的程序被处理器执行时可实现如上述的自动驾驶控制方法的步骤。The present application also provides an electronic device, which is a physical device, and includes: a memory, a processor, and a program of the autonomous driving control method stored in the memory and executable on the processor. When the program of the autonomous driving control method is executed by the processor, the steps of the autonomous driving control method as described above can be implemented.
本申请还提供一种存储介质,所述存储介质为计算机可读存储介质,所述计算机可读存储介质上存储有实现自动驾驶控制方法的程序,所述的自动驾驶控制方法的程序被处理器执行时实现如上述的自动驾驶控制方法的步骤。The present application also provides a storage medium, which is a computer-readable storage medium. The computer-readable storage medium stores a program for implementing an autonomous driving control method. When the program of the autonomous driving control method is executed by a processor, the steps of the autonomous driving control method as described above are implemented.
本申请提供了一种自动驾驶控制方法、装置、电子设备及存储介质,所述自动驾驶控制方法采用认知决策模型,所述认知决策模型包括认知层和决策层。首先,通过获取车内外监测数据,通过所述认知层从所述车内外监测数据中提取当前认知层特征,实现了对自动驾驶车辆当前情况的认知;进而通过将所述环境认知特征和所述用户需求特征输入所述决策层,确定自动驾驶控制参数,实现了在认知到当前情况的基础上,准确作出自动驾驶决策,确定自动驾驶控制参数;进而通过根据所述当前认知层特征和所述自动驾驶控制参数,生成驾驶行为解释,实现对自动驾驶决策进行解释的目的。由于自动驾驶控制参数是基于认知层特征进行决策确定的,因此自动驾驶控制参数与认知层特征之间存在一定的因果关系,而用户之所以无法理解自动驾驶车辆的驾驶行为,正是因为用户在感受到自动驾驶车辆突发性的驾驶行为之前,注意力可能与驾驶车辆的注意力集中在不同的事务上,因此在感受到自动驾驶车辆突发性的驾驶行为之时,感受到的是自动驾驶车辆最终作出的决策,可能看不到影响决策的因素,从而无法推理出其作出决策的原因,因此会产生疑惑、恐慌、不安等负面情绪。因此,本申请基于认知层特征对自动驾驶控制参数进行解释,生成驾驶行为解释,可以主动与用户进行交互,为用户解释驾驶行为产生的原因,消除用户对驾驶行为的疑惑。因此克服了端到端感知决策架构的可解释性较差,因此在控制车辆执行突发性行为时,容易导致乘坐自动驾驶车辆的用户无法理解自动驾驶车辆的驾驶行为而产生疑惑、恐慌、不安等负面情绪,用户体验感较差的技术缺陷,提高了端到端感知决策架构的可解释性,减少了用户因未知而产生的疑惑、恐慌、不安等负面情绪,提高了用户体验感。The present application provides an automatic driving control method, device, electronic device and storage medium, wherein the automatic driving control method adopts a cognitive decision model, wherein the cognitive decision model includes a cognitive layer and a decision layer. First, by acquiring the monitoring data inside and outside the vehicle, the current cognitive layer features are extracted from the monitoring data inside and outside the vehicle through the cognitive layer, thereby realizing the recognition of the current situation of the automatic driving vehicle; then, by inputting the environmental cognitive features and the user demand features into the decision layer, the automatic driving control parameters are determined, thereby realizing accurate automatic driving decisions and determining automatic driving control parameters based on the recognition of the current situation; then, by generating a driving behavior explanation based on the current cognitive layer features and the automatic driving control parameters, the purpose of explaining the automatic driving decision is realized. Since the automatic driving control parameters are determined based on the cognitive layer features, there is a certain causal relationship between the automatic driving control parameters and the cognitive layer features, and the reason why the user cannot understand the driving behavior of the automatic driving vehicle is that before the user feels the sudden driving behavior of the automatic driving vehicle, the user's attention may be focused on different matters from the driving vehicle. Therefore, when the user feels the sudden driving behavior of the automatic driving vehicle, what he feels is the final decision made by the automatic driving vehicle, and he may not see the factors affecting the decision, and thus cannot infer the reason for its decision, so he will have negative emotions such as doubt, panic, and anxiety. Therefore, this application interprets the autonomous driving control parameters based on the cognitive layer characteristics, generates a driving behavior explanation, can actively interact with the user, explain the reasons for the driving behavior to the user, and eliminate the user's doubts about the driving behavior. Therefore, it overcomes the poor interpretability of the end-to-end perception decision architecture. Therefore, when controlling the vehicle to perform sudden behaviors, it is easy to cause users riding in the autonomous driving vehicle to be unable to understand the driving behavior of the autonomous driving vehicle and generate negative emotions such as doubt, panic, and uneasiness, and the technical defects of poor user experience. It improves the interpretability of the end-to-end perception decision architecture, reduces the user's negative emotions such as doubt, panic, and uneasiness caused by the unknown, and improves the user experience.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
此处的附图被并入说明书中并构成本说明书的一部分,示出了符合本申请的实施例,并与说明书一起用于解释本申请的原理。The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present application and, together with the description, serve to explain the principles of the present application.
为了更清楚地说明本申请实施例或相关技术中的技术方案,下面将对实施例或相关技术描述中所需要使用的附图作简单地介绍,显而易见地,对于本领域普通技术人员而言,在不付出创造性劳动性的前提下,还可以根据这些附图得到其他的附图。In order to more clearly illustrate the technical solutions in the embodiments of the present application or the related technologies, the drawings required for use in the embodiments or the related technical descriptions are briefly introduced below. Obviously, for ordinary technicians in this field, other drawings can be obtained based on these drawings without paying any creative labor.
图1为本申请自动驾驶控制方法的第一实施例的流程示意图;FIG1 is a schematic flow chart of a first embodiment of an automatic driving control method of the present application;
图2为本申请实施例中认知决策模型的一种可实施方式的结构示意图;FIG2 is a schematic diagram of a structure of a possible implementation method of a cognitive decision-making model in an embodiment of the present application;
图3为本申请自动驾驶控制方法的第二实施例的流程示意图FIG. 3 is a flow chart of a second embodiment of the automatic driving control method of the present application.
图4为本申请自动驾驶控制方法的第三实施例的流程示意图;FIG4 is a schematic diagram of a flow chart of a third embodiment of the automatic driving control method of the present application;
图5为本申请实施例中自动驾驶控制装置的结构示意图;FIG5 is a schematic diagram of the structure of an automatic driving control device in an embodiment of the present application;
图6为本申请实施例中自动驾驶控制方法涉及的硬件运行环境的设备结构示意图。FIG6 is a schematic diagram of the device structure of the hardware operating environment involved in the automatic driving control method in an embodiment of the present application.
本申请目的实现、功能特点及优点将结合实施例,参照附图做进一步说明。The purpose, features and advantages of this application will be further described in conjunction with the embodiments and with reference to the accompanying drawings.
具体实施方式DETAILED DESCRIPTION
为使本发明的上述目的、特征和优点能够更加明显易懂,下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述。显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有作出创造性劳动的前提下所得到的所有其它实施例,均属于本发明保护的范围。In order to make the above-mentioned purposes, features and advantages of the present invention more obvious and easy to understand, the technical scheme in the embodiments of the present invention will be clearly and completely described below in conjunction with the drawings in the embodiments of the present invention. Obviously, the described embodiments are only part of the embodiments of the present invention, not all of the embodiments. Based on the embodiments of the present invention, all other embodiments obtained by ordinary technicians in this field without making creative work belong to the scope of protection of the present invention.
实施例一Embodiment 1
本申请实施例提供一种自动驾驶控制方法,参照图1,在本申请自动驾驶控制方法的第一实施例中,所述自动驾驶控制方法采用认知决策模型,所述认知决策模型包括认知层和决策层,所述自动驾驶控制方法包括以下步骤:The present application provides an automatic driving control method. Referring to FIG. 1 , in the first embodiment of the automatic driving control method of the present application, the automatic driving control method adopts a cognitive decision model, and the cognitive decision model includes a cognitive layer and a decision layer. The automatic driving control method includes the following steps:
步骤S10,获取车内外监测数据,通过所述认知层从所述车内外监测数据中提取当前认知层特征;Step S10, acquiring vehicle interior and exterior monitoring data, and extracting current cognitive layer features from the vehicle interior and exterior monitoring data through the cognitive layer;
本实施例方法的执行主体可以是一种自动驾驶控制装置,也可以是一种自动驾驶控制终端设备或服务器,本实施例以自动驾驶控制装置进行举例,该自动驾驶控制装置可以集成在具有数据处理功能的车辆、车载终端、智能手机、计算机等终端设备上。The executor of the method of this embodiment can be an autonomous driving control device, or an autonomous driving control terminal device or server. This embodiment takes an autonomous driving control device as an example. The autonomous driving control device can be integrated in a terminal device such as a vehicle with data processing function, an on-board terminal, a smart phone, a computer, etc.
在本实施例中,需要说明的是,所述自动驾驶控制方法应用于认知决策模型,所述认知决策模型用于实现对自动驾驶车辆进行端到端的驾驶行为控制,所述认知决策模型使用注意力机制,可以为基于Transformer结构的深度学习模型,也可以为其他使用注意力机制的深度学习神经网络模型,具体可以根据实际情况进行设计,本实施例对此不加以限制。In this embodiment, it should be noted that the autonomous driving control method is applied to a cognitive decision-making model, which is used to achieve end-to-end driving behavior control of the autonomous driving vehicle. The cognitive decision-making model uses an attention mechanism and can be a deep learning model based on a Transformer structure or other deep learning neural network models using an attention mechanism. The specific design can be based on actual conditions, and this embodiment does not impose any restrictions on this.
所述认知决策模型至少包括认知层和决策层,其中,认知层用于对获取到的车内外监测数据进行特征提取、融合、编码和推理等得到决策层所需的风险特征;所述决策层用于基于认知层传递的风险特征规划自动驾驶车辆在下一时间步的驾驶行为,以使得自动驾驶车辆可以规避风险安全行驶到目的地。The cognitive decision model includes at least a cognitive layer and a decision layer, wherein the cognitive layer is used to extract features, fuse, encode and infer the acquired in-vehicle and out-vehicle monitoring data to obtain the risk features required by the decision layer; the decision layer is used to plan the driving behavior of the autonomous driving vehicle at the next time step based on the risk features transmitted by the cognitive layer, so that the autonomous driving vehicle can avoid risks and drive safely to the destination.
所述车内外监测数据是指能够获取到的自动驾驶控制所需的数据,包括车外监测数据和座舱监测数据,所述车内外监测数据可以为图像数据、文本数据、声音数据、信号数据等中的一种或多种模态。示例性地,所述车外监测数据可以为通过摄像头采集的车外图像数据、通过麦克风采集的车外声音数据、通过通信模块获取到的信号灯读秒数据等。示例性地,所述座舱监测数据可以为从整车控制器或控制器局域网总线上获取的车速、加速度等车辆行驶数据,也可以为通过摄像头采集的座舱图像数据、驾驶员图像数据、通过麦克风采集的座舱声音数据、将采集到的语音转换得到的文本数据、从采集到的图像中提取的文本数据等。这些车内外监测数据可以实时采集,因此可以携带有其他交通参与者的状态、驾驶员的状态、路况状态等可能发生变化的信息在当前时刻的实时信息。The in-vehicle and out-vehicle monitoring data refers to the data required for automatic driving control that can be obtained, including in-vehicle monitoring data and cockpit monitoring data. The in-vehicle and out-vehicle monitoring data can be one or more modes of image data, text data, sound data, signal data, etc. Exemplarily, the in-vehicle monitoring data can be in-vehicle image data collected by a camera, in-vehicle sound data collected by a microphone, signal light countdown data obtained by a communication module, etc. Exemplarily, the cockpit monitoring data can be vehicle driving data such as vehicle speed and acceleration obtained from a vehicle controller or a controller area network bus, or can be cockpit image data collected by a camera, driver image data, cockpit sound data collected by a microphone, text data obtained by converting the collected voice, text data extracted from the collected image, etc. These in-vehicle and out-vehicle monitoring data can be collected in real time, so they can carry real-time information at the current moment about the status of other traffic participants, the status of the driver, the road condition, and other information that may change.
对于不同模态的车内外监测数据,可以采用不同的模型进行特征提取,例如,图像数据可以采用卷积神经网络进行特征提取,时序数据可以采用长短期记忆网络进行特征提取等,也可以采用不同的模型进行特征处理从而提取出决策层所需的不同的风险特征,因此,认知层中可以部署多个模型进行特征处理,具体的模型结构和模型类型可以根据实际情况进行确定,本实施例对此不加以限制。Different models can be used to extract features from in-vehicle and out-of-vehicle monitoring data of different modalities. For example, convolutional neural networks can be used to extract features from image data, and long short-term memory networks can be used to extract features from time series data. Different models can also be used for feature processing to extract different risk features required by the decision-making layer. Therefore, multiple models can be deployed in the cognitive layer for feature processing. The specific model structure and model type can be determined based on actual conditions, and this embodiment does not impose any restrictions on this.
所述认知层特征可以包括环境认知特征和用户需求特征等中的至少一种。The cognitive layer features may include at least one of environmental cognitive features and user demand features.
所述环境认知特征用于表征自动驾驶车辆对当前所处环境的认知,例如对于交通设施的认知、对于地面障碍物的认知、对于本车行驶状态的认知、对于其他交通参与者的认知、对于所处场景的风险的认知、对于所处场景的风险项的认知等。自动驾驶车辆行为控制的主要目的是使得自动驾驶车辆安全到达目的地,也即,在规避风险的情况下控制车辆行驶到目的地,在一种可实施的方式中,自动驾驶控制的方式可以为预测自动驾驶车辆在下一时间步的高风险行驶区域和/或低风险行驶区域,从而规划出自动驾驶车辆在下一时间步避开高风险行驶区域或经过低风险行驶区域的驾驶行为。而所述环境认知特征是自动驾驶车辆当前所处环境的客观表示,因此提取到的所述环境认知特征越准确,对于自动驾驶车辆当前的风险判断越准确,自动驾驶控制的安全性则越高。所述环境认知特征可以包括场景认知特征、地图特征、风险目标检测跟踪特征、风险目标运动特征、低风险行驶区域特征等,还可以包括其他特征,所述认知层可以为多层结构,也即,其中任意一层认知层的输入都可以为其他认知层的输出,任意一层认知层的输出也都可以作为其他认知层的输入,从而可以实现多特征的融合和推理,捕捉认知层提取的各个特征之间的关联性和依赖性,实现对当前驾驶实际情况进行更全面的理解和更深层的逻辑推理。The environmental cognition feature is used to characterize the cognition of the autonomous driving vehicle to the current environment, such as the cognition of traffic facilities, ground obstacles, the driving status of the vehicle, other traffic participants, the risk of the scene, the risk items of the scene, etc. The main purpose of the behavior control of the autonomous driving vehicle is to make the autonomous driving vehicle reach the destination safely, that is, to control the vehicle to drive to the destination while avoiding risks. In an implementable manner, the autonomous driving control method can be to predict the high-risk driving area and/or low-risk driving area of the autonomous driving vehicle in the next time step, so as to plan the driving behavior of the autonomous driving vehicle to avoid the high-risk driving area or pass through the low-risk driving area in the next time step. The environmental cognition feature is an objective representation of the current environment of the autonomous driving vehicle. Therefore, the more accurate the extracted environmental cognition feature is, the more accurate the current risk judgment of the autonomous driving vehicle is, and the higher the safety of the autonomous driving control is. The environmental cognition features may include scene cognition features, map features, risk target detection and tracking features, risk target motion features, low-risk driving area features, etc., and may also include other features. The cognitive layer may be a multi-layer structure, that is, the input of any cognitive layer can be the output of other cognitive layers, and the output of any cognitive layer can also be used as the input of other cognitive layers, thereby realizing the fusion and reasoning of multiple features, capturing the correlation and dependency between the various features extracted by the cognitive layer, and achieving a more comprehensive understanding of the current actual driving situation and deeper logical reasoning.
所述用户需求特征用于表征自动驾驶车辆上乘坐的用户的需求,例如,用户对于目的地的需求、用户对于行驶到目的地的路径的需求、用户赶时间的需求、用户希望平缓行驶减缓晕车的需求等。所述用户需求特征可以通过检测用户操作、采集用户图像等方式捕捉用户需求信息,进而从中提取出用户需求特征,例如,可以通过摄像头采集用户图像,在通过图像识别技术识别出用户图像中用户当前处于睡眠状态的情况下,用户在睡眠状态通常隐含了用户对车辆平稳行驶的需求,例如,还可以通过麦克风采集用户语音信息,在识别出用户说“开快一点”的情况下,可以知道用户有加速的需求。The user demand feature is used to characterize the needs of users riding in the autonomous driving vehicle, such as the user's needs for the destination, the user's needs for the route to the destination, the user's needs for being in a hurry, the user's needs for smooth driving to relieve motion sickness, etc. The user demand feature can capture user demand information by detecting user operations, collecting user images, etc., and then extract user demand features therefrom. For example, the user image can be collected by a camera. When the image recognition technology is used to identify that the user is currently in a sleeping state in the user image, the user's sleeping state usually implies the user's need for the vehicle to drive smoothly. For example, the user's voice information can also be collected by a microphone. When it is recognized that the user says "drive faster", it can be known that the user has a need to accelerate.
所述认知层可以采用注意力机制,在一种可实施的方式中,所述认知层可以采用transformer结构。采用注意力机制可以很好地对各种模态各种类型的车辆监测数据进行特征提取、融合、编码和推理,提取出可能发生变化的信息在当前时刻的实时信息,与当前时刻之前一段历史时间的历史信息组成信息序列,并捕捉不同类型的信息之间的关联性和依赖性,进行特征推理,提取出具有深层逻辑信息的风险特征,以供决策层对自动驾驶车辆进行自动驾驶控制,这样,可以在预测自动驾驶车辆行车风险的过程中充分考虑到可能发生变化的信息、多个风险项以及用户需求之间的相互影响,提高自动驾驶车辆行车风险的预测准确性,从而使得自动驾驶车辆可以规避更多的风险,同时更好地满足用户需求,提高自动驾驶控制的安全性和舒适性。其中,不同类型的信息之间的关联性和依赖性可能会对自动驾驶车辆的自动驾驶控制有较大影响,示例性地,交通参与者与交通参与者之间可能产生交互、交通参与者与交通设施之间可能产生交互、交通参与者与场景之间也可能产生交互。例如,十字路口处,本车为东西方向行驶,附近还有南北方向行驶的第二车辆和东西方向行走的行人,假设在不考虑交通参与者与交通参与者之间产生交互的情况下,本车和第二车辆以当前车速行驶不会发生碰撞,但若第二车辆因为避让行人而减速或停车,则可能导致本车与第二车辆发生碰撞,因此,在考虑交通参与者与交通参与者之间产生交互的情况下,本车仍需要减速或者规划其他行驶路线;而若假设第二车辆位于转弯车道,在考虑交通参与者与交通设施之间产生交互的情况下,若第二车辆转弯的行驶轨迹与本车的行驶轨迹不相交,本车则无需减速;另外,在考虑交通参与者与场景之间产生交互的情况下,在当前处于经过十字路口的场景下,由于十字路口场景的复杂性,可以提高十字路口处人和车的风险系数,因此可能需要控制车辆减速慢行。The cognitive layer can adopt an attention mechanism. In an implementable manner, the cognitive layer can adopt a transformer structure. The attention mechanism can be used to perform feature extraction, fusion, encoding and reasoning on various types of vehicle monitoring data of various modalities, extract the real-time information of the information that may change at the current moment, and form an information sequence with the historical information of a period of time before the current moment, and capture the correlation and dependency between different types of information, perform feature reasoning, and extract risk features with deep logical information for the decision-making layer to perform autonomous driving control on the autonomous driving vehicle. In this way, the mutual influence between the information that may change, multiple risk items and user needs can be fully considered in the process of predicting the driving risk of the autonomous driving vehicle, and the prediction accuracy of the driving risk of the autonomous driving vehicle can be improved, so that the autonomous driving vehicle can avoid more risks, and at the same time better meet user needs and improve the safety and comfort of autonomous driving control. Among them, the correlation and dependency between different types of information may have a greater impact on the autonomous driving control of the autonomous driving vehicle. For example, traffic participants may interact with each other, traffic participants may interact with traffic facilities, and traffic participants may interact with scenes. For example, at an intersection, the vehicle is traveling in the east-west direction, and there is a second vehicle traveling in the north-south direction and pedestrians walking in the east-west direction nearby. Assuming that the interaction between traffic participants is not considered, the vehicle and the second vehicle will not collide at the current speed. However, if the second vehicle slows down or stops to avoid pedestrians, it may cause a collision between the vehicle and the second vehicle. Therefore, considering the interaction between traffic participants, the vehicle still needs to slow down or plan other driving routes; and if it is assumed that the second vehicle is in a turning lane, considering the interaction between traffic participants and traffic facilities, if the turning trajectory of the second vehicle does not intersect with the driving trajectory of the vehicle, the vehicle does not need to slow down; in addition, considering the interaction between traffic participants and the scene, in the current scenario of passing through an intersection, due to the complexity of the intersection scene, the risk factor of people and vehicles at the intersection can be increased, so it may be necessary to control the vehicle to slow down.
作为一种示例,所述步骤S10包括:通过设置于自动驾驶车辆上的一个或多个传感器,采集当前时刻的传感器数据作为车内外监测数据,还可以通过自动驾驶车辆上的通讯模块,与外部设备进行通信,获取外部设备提供的外部数据作为车内外监测数据,所述外部数据例如红绿灯信号、交通事故信息、路侧监测设备采集的路侧监测数据等。将所述车内外监测数据输入所述认知层,进行特征提取、特征融合、特征编码、特征推理等多层次递进式的特征处理,得到认知层特征。As an example, step S10 includes: collecting sensor data at the current moment as vehicle interior and exterior monitoring data through one or more sensors provided on the autonomous driving vehicle, and also communicating with external devices through a communication module on the autonomous driving vehicle to obtain external data provided by the external devices as vehicle interior and exterior monitoring data, such as traffic light signals, traffic accident information, and roadside monitoring data collected by roadside monitoring equipment. The vehicle interior and exterior monitoring data are input into the cognitive layer, and multi-level progressive feature processing such as feature extraction, feature fusion, feature encoding, and feature reasoning is performed to obtain cognitive layer features.
可选地,所述认知层包括用户特征提取模型;所述当前认知层特征包括用户需求特征,其中,所述用户需求特征包括用户隐式需求特征和用户显式需求特征;所述车内外监测数据包括用户操作数据和用户状态数据;Optionally, the cognitive layer includes a user feature extraction model; the current cognitive layer features include user demand features, wherein the user demand features include user implicit demand features and user explicit demand features; the in-vehicle and out-of-vehicle monitoring data include user operation data and user status data;
所述通过所述认知层从所述车内外监测数据中提取当前认知层特征的步骤包括:The step of extracting current cognitive layer features from the vehicle internal and external monitoring data through the cognitive layer includes:
通过所述用户特征提取模型从所述用户操作数据中提取用户显示需求特征,和/或,通过所述用户特征提取模型从所述用户状态数据中提取用户隐式需求特征。The user feature extraction model is used to extract user explicit demand features from the user operation data, and/or the user feature extraction model is used to extract user implicit demand features from the user status data.
在本实施例中,需要说明的是,用户操作数据是指用户主动与自动驾驶车辆进行交互而产生的数据,例如用户对自动驾驶车辆发出语音指令,或主动点击车载终端显示界面上的按键等等。用户状态数据是指自动驾驶车辆主动采集的用户相关的信息,例如在用户睡着时,用户没有与自动驾驶车辆主动进行交互,此时可以通过摄像头采集用户图像数据,识别出用户睡着的状态,又例如,用户在和朋友打电话表达出乘坐自动驾驶车辆感到紧张和忐忑的状态,用户没有与自动驾驶车辆主动进行交互,此时可以通过麦克风采集用户语音数据,识别出用户紧张和忐忑的状态。In this embodiment, it should be noted that user operation data refers to data generated by users actively interacting with the autonomous vehicle, such as the user issuing voice commands to the autonomous vehicle, or actively clicking buttons on the display interface of the vehicle terminal, etc. User status data refers to user-related information actively collected by the autonomous vehicle. For example, when the user is asleep, the user does not actively interact with the autonomous vehicle. At this time, the user's image data can be collected through the camera to identify the user's sleeping state. For another example, when the user is calling a friend to express the state of nervousness and anxiety about riding in the autonomous vehicle, the user does not actively interact with the autonomous vehicle. At this time, the user's voice data can be collected through the microphone to identify the user's nervousness and anxiety.
作为一种示例,可以将所述用户操作数据输入所述用户特征提取模型,进行特征提取、特征融合、特征编码、特征推理等多层次递进式的特征处理,得到用户显示需求特征,也可以将所述用户状态数据输入所述用户特征提取模型,进行特征提取、特征融合、特征编码、特征推理等多层次递进式的特征处理,得到用户隐式需求特征。As an example, the user operation data can be input into the user feature extraction model to perform multi-level progressive feature processing such as feature extraction, feature fusion, feature encoding, and feature reasoning to obtain user display demand features. The user status data can also be input into the user feature extraction model to perform multi-level progressive feature processing such as feature extraction, feature fusion, feature encoding, and feature reasoning to obtain user implicit demand features.
在本实施例中,用户可以主动与自动驾驶车辆进行交互,主动提出需求,自动驾驶车辆也可以响应用户的交互操作,满足用户需求,提高自动驾驶车辆与用户之间的交互性,自动驾驶车辆还可以主动发现用户需求,主动作出更符合用户需求的自动驾驶行为,从而减少用户产生不适感。In this embodiment, the user can actively interact with the autonomous driving vehicle and actively put forward demands. The autonomous driving vehicle can also respond to the user's interactive operations, meet the user's needs, and improve the interactivity between the autonomous driving vehicle and the user. The autonomous driving vehicle can also actively discover user needs and actively make autonomous driving behaviors that better meet user needs, thereby reducing user discomfort.
可选地,所述认知层包括场景认知模型、地图构建模型、风险目标检测跟踪模型和运动预测模型;所述当前认知层特征包括环境认知特征,其中,所述环境认知特征包括场景认知特征、地图特征、风险目标检测跟踪特征和风险目标运动特征;Optionally, the cognitive layer includes a scene cognitive model, a map construction model, a risk target detection and tracking model, and a motion prediction model; the current cognitive layer features include environmental cognitive features, wherein the environmental cognitive features include scene cognitive features, map features, risk target detection and tracking features, and risk target motion features;
所述通过所述认知层从所述车内外监测数据中提取当前认知层特征的步骤包括:The step of extracting current cognitive layer features from the vehicle internal and external monitoring data through the cognitive layer includes:
步骤A10,通过所述场景认知模型从所述车内外监测数据中提取出场景认知特征,通过所述地图构建模型从所述车内外监测数据中提取出地图特征,并通过所述风险目标检测跟踪模型从所述车内外监测数据中提取出风险目标检测跟踪特征;Step A10, extracting scene recognition features from the in-vehicle and out-vehicle monitoring data by using the scene recognition model, extracting map features from the in-vehicle and out-vehicle monitoring data by using the map building model, and extracting risk target detection and tracking features from the in-vehicle and out-vehicle monitoring data by using the risk target detection and tracking model;
步骤A20,通过所述运动预测模型采用注意力机制,基于所述场景认知特征、所述地图特征和所述风险目标检测跟踪特征,进行风险目标运动预测,得到风险目标运动特征。Step A20, using the attention mechanism through the motion prediction model, based on the scene recognition features, the map features and the risk target detection and tracking features, performs risk target motion prediction to obtain risk target motion features.
在本实施例中,需要说明的是,所述认知层包括场景认知模型、地图构建模型、风险目标检测跟踪模型和运动预测模型,其中,风险目标是指对自动驾驶车辆行驶可能造成风险的物体,可以为交通参与者、交通基础设施等中的一种或多种。所述场景认知模型、所述地图构建模型和所述风险目标检测跟踪模型输出的场景认知特征、地图特征、风险目标检测跟踪特征,均对预测风险目标的在下一时间步的运动有一定影响,且场景认知特征、地图特征和风险目标检测跟踪特征之间还存在一定的关联性和依赖性,因此均可以作为运动预测模型的输入,通过运动预测模型采用注意力机制,捕捉这些特征之间的关联性和依赖性,进行进一步的特征融合、编码和推理,确定风险目标的风险目标运动特征。In this embodiment, it should be noted that the cognitive layer includes a scene cognitive model, a map construction model, a risk target detection and tracking model, and a motion prediction model, wherein a risk target refers to an object that may cause risk to the driving of an autonomous vehicle, and may be one or more of a traffic participant, a traffic infrastructure, etc. The scene cognitive features, map features, and risk target detection and tracking features output by the scene cognitive model, the map construction model, and the risk target detection and tracking model all have a certain influence on predicting the motion of the risk target at the next time step, and there is a certain correlation and dependency between the scene cognitive features, the map features, and the risk target detection and tracking features, so they can all be used as inputs to the motion prediction model. The motion prediction model uses an attention mechanism to capture the correlation and dependency between these features, and further feature fusion, encoding, and reasoning are performed to determine the risk target motion features of the risk target.
所述风险特征包括场景认知特征、地图特征、风险目标检测跟踪特征和风险目标运动特征。场景认知特征用于表征对车辆当前所处场景的风险认知,例如注意力分布、对场景中存在的风险项的语义理解等等,地图特征用于表征车辆所处场景的地面设施、道路元素等地图信息,风险目标检测跟踪特征用于表征风险目标的识别以及当前时刻之前的一段时间范围内风险目标的位置信息,风险目标运动特征用于表征风险目标在当前时刻之后的一段时间范围的运动信息,运动信息可以包括位置信息、轨迹信息、运动意图、运动趋势等。The risk features include scene recognition features, map features, risk target detection and tracking features, and risk target motion features. Scene recognition features are used to characterize the risk recognition of the scene in which the vehicle is currently located, such as attention distribution, semantic understanding of risk items in the scene, etc. Map features are used to characterize map information such as ground facilities and road elements in the scene in which the vehicle is located. Risk target detection and tracking features are used to characterize the identification of risk targets and the location information of risk targets within a period of time before the current moment. Risk target motion features are used to characterize the motion information of risk targets within a period of time after the current moment. The motion information may include location information, trajectory information, motion intention, motion trend, etc.
作为一种示例,所述步骤A10-A20包括:将所述车内外监测数据输入所述场景认知模型,进行特征提取、特征融合、特征编码、特征推理等多层次递进式的特征处理,提取出场景认知特征;同步地,将所述车内外监测数据输入所述地图构建模型,进行特征提取、特征融合、特征编码、特征推理等多层次递进式的特征处理,提取出地图特征;同步地,将所述车内外监测数据输入所述风险目标检测跟踪模型,进行特征提取、特征融合、特征编码、特征推理等多层次递进式的特征处理,提取出风险目标检测跟踪特征。进而,将所述场景认知特征、所述地图特征和所述风险目标检测跟踪特征输入所述运动预测模型,通过运动预测模型,采用注意力机制,捕捉所述场景认知特征、所述地图特征和所述风险目标检测跟踪特征之间的关联性和依赖性,基于这些关联性和依赖性,可以更准确地进行风险目标运行预测,确定至少一个风险目标的风险目标运动特征。As an example, the steps A10-A20 include: inputting the vehicle interior and exterior monitoring data into the scene recognition model, performing multi-level progressive feature processing such as feature extraction, feature fusion, feature coding, and feature reasoning, and extracting scene recognition features; synchronously, inputting the vehicle interior and exterior monitoring data into the map construction model, performing multi-level progressive feature processing such as feature extraction, feature fusion, feature coding, and feature reasoning, and extracting map features; synchronously, inputting the vehicle interior and exterior monitoring data into the risk target detection and tracking model, performing multi-level progressive feature processing such as feature extraction, feature fusion, feature coding, and feature reasoning, and extracting risk target detection and tracking features. Further, the scene recognition features, the map features, and the risk target detection and tracking features are input into the motion prediction model, and the motion prediction model is used to capture the correlation and dependency between the scene recognition features, the map features, and the risk target detection and tracking features by using the attention mechanism. Based on these correlations and dependencies, risk target operation prediction can be performed more accurately, and the risk target motion features of at least one risk target can be determined.
在一种可实施的方式中,所述风险目标检测跟踪特征包括至少一个风险目标检测跟踪子特征;所述运动预测模型包括注意力层、多层感知机层和预测层;In an implementable manner, the risk target detection and tracking feature includes at least one risk target detection and tracking sub-feature; the motion prediction model includes an attention layer, a multi-layer perceptron layer, and a prediction layer;
所述通过所述运动预测模型采用注意力机制,基于所述场景认知特征、所述地图特征和所述风险目标检测跟踪特征,进行风险目标运动预测,得到风险目标运动特征的步骤包括:The step of using the attention mechanism through the motion prediction model to predict the motion of the risk target based on the scene recognition features, the map features and the risk target detection and tracking features to obtain the motion features of the risk target includes:
步骤A21,将所述场景认知特征、所述地图特征和所述风险目标检测跟踪特征输入所述注意力层,采用注意力机制,确定各所述风险目标检测跟踪子特征之间的风险目标交互特征、各所述风险目标检测跟踪子特征与所述场景认知特征之间的场景交互特征以及各所述风险目标检测跟踪子特征与所述地图特征之间的地图交互特征;Step A21, inputting the scene recognition feature, the map feature and the risk target detection and tracking feature into the attention layer, using the attention mechanism to determine the risk target interaction feature between each of the risk target detection and tracking sub-features, the scene interaction feature between each of the risk target detection and tracking sub-features and the scene recognition feature, and the map interaction feature between each of the risk target detection and tracking sub-features and the map feature;
步骤A22,将所述风险目标交互特征、所述场景交互特征以及所述地图交互特征输入多层感知机层,得到场景风险特征和运动查询特征;Step A22, inputting the risk target interaction feature, the scene interaction feature and the map interaction feature into a multi-layer perceptron layer to obtain a scene risk feature and a motion query feature;
步骤A23,将所述场景风险特征和所述运动查询特征输入所述预测层,进行风险目标运动预测,得到风险目标运动特征。Step A23, inputting the scene risk feature and the motion query feature into the prediction layer, performing risk target motion prediction, and obtaining risk target motion features.
在本实施例中,需要说明的是,对于复杂场景,可能涉及多个风险目标,在识别到多个风险目标的情况下,风险目标检测跟踪模型可以识别出多个风险目标并对多个风险目标同步进行跟踪,这样获得的风险目标检测跟踪特征中就会包含有多个风险目标的相关信息,也即,所述风险目标检测跟踪特征包括至少一个风险目标检测跟踪子特征。在此情况下,可以对每个风险目标分别进行运动预测。通过采用注意力机制,可以捕捉每个风险目标与其他风险目标之间的关联性和依赖性、每个风险目标与地图元素之间的关联性和依赖性、每个风险目标与场景认知之间的关联性和依赖性,基于这些关联性和依赖性,可以更准确地对每个风险目标进行风险目标运行预测,确定每个风险目标的风险目标运动特征。In this embodiment, it should be noted that for complex scenarios, multiple risk targets may be involved. When multiple risk targets are identified, the risk target detection and tracking model can identify multiple risk targets and track multiple risk targets simultaneously, so that the risk target detection and tracking features obtained will contain relevant information of multiple risk targets, that is, the risk target detection and tracking features include at least one risk target detection and tracking sub-feature. In this case, motion prediction can be performed for each risk target separately. By adopting the attention mechanism, the correlation and dependency between each risk target and other risk targets, the correlation and dependency between each risk target and map elements, and the correlation and dependency between each risk target and scene cognition can be captured. Based on these correlations and dependencies, risk target operation prediction can be performed more accurately for each risk target, and the risk target motion characteristics of each risk target can be determined.
场景风险特征和运动查询特征表征的是当前场景已经存在的风险,基于场景风险特征和运动查询特征可以进一步预测风险目标在当前时刻之后的运动信息。The scene risk features and motion query features represent the risks that already exist in the current scene. Based on the scene risk features and motion query features, the motion information of the risk target after the current moment can be further predicted.
作为一种示例,所述步骤A21-A23包括:将所述场景认知特征、所述地图特征和所述风险目标检测跟踪特征输入所述注意力层,采用注意力机制,捕捉各所述风险目标检测跟踪子特征之间的依赖关系,以确定各个风险目标与其他风险目标之间的相互影响,生成风险目标交互特征,捕捉每个所述风险目标检测跟踪子特征与所述场景认知特征之间的依赖关系,以确定每个风险目标与场景认知之间的相互影响,生成场景交互特征,捕捉每个所述风险目标检测跟踪子特征与所述地图特征之间的依赖关系,以确定每个风险目标与地图元素之间的相互影响,生成场景交互特征。进而,将所述风险目标交互特征、所述场景交互特征以及所述地图交互特征拼接后,输入多层感知机层,得到场景风险特征和运动查询特征。进而,将所述场景风险特征和所述运动查询特征输入所述预测层,对每个风险目标进行风险目标运动预测,得到每个风险目标的风险目标运动特征。As an example, the steps A21-A23 include: inputting the scene recognition feature, the map feature and the risk target detection and tracking feature into the attention layer, using the attention mechanism to capture the dependency between each of the risk target detection and tracking sub-features to determine the mutual influence between each risk target and other risk targets, generating a risk target interaction feature, capturing the dependency between each of the risk target detection and tracking sub-features and the scene recognition feature to determine the mutual influence between each risk target and scene recognition, generating a scene interaction feature, capturing the dependency between each of the risk target detection and tracking sub-features and the map feature to determine the mutual influence between each risk target and the map element, generating a scene interaction feature. Then, the risk target interaction feature, the scene interaction feature and the map interaction feature are spliced and input into the multi-layer perceptron layer to obtain the scene risk feature and the motion query feature. Then, the scene risk feature and the motion query feature are input into the prediction layer to perform risk target motion prediction on each risk target to obtain the risk target motion feature of each risk target.
在一种可实施的方式中,多层感知机层还可以解码场景风险特征和运动查询特征,将解码得到的场景风险预测信息和风险目标的运动信息作为中间结果输出,以供用户查看。In one feasible manner, the multi-layer perceptron layer can also decode scene risk features and motion query features, and output the decoded scene risk prediction information and motion information of the risk target as intermediate results for user viewing.
步骤S20,将所述当前认知层特征输入所述决策层,确定自动驾驶控制参数;Step S20, inputting the current cognitive layer features into the decision layer to determine the automatic driving control parameters;
在本实施例中,需要说明的是,所述自动驾驶控制参数可以包括行驶轨迹、车辆控制参数等,其中,行驶轨迹是指自动驾驶车辆在当前时刻之后一段时间的轨迹,可以控制自动驾驶车辆基于所述行驶轨迹行驶,所述车辆控制参数包括速度、油门踏板开度、方向盘角度等。In this embodiment, it should be noted that the autonomous driving control parameters may include driving trajectory, vehicle control parameters, etc., wherein the driving trajectory refers to the trajectory of the autonomous driving vehicle a period of time after the current moment, and the autonomous driving vehicle can be controlled to drive based on the driving trajectory. The vehicle control parameters include speed, accelerator pedal opening, steering wheel angle, etc.
作为一种示例,所述步骤S20包括:将所述环境认知特征和所述用户需求特征输入所述决策层,确定自动驾驶车辆在当前时刻或当前时刻之后一段时间内的自动驾驶控制参数。As an example, the step S20 includes: inputting the environmental cognition characteristics and the user demand characteristics into the decision layer to determine the autonomous driving control parameters of the autonomous driving vehicle at the current moment or within a period of time after the current moment.
步骤S30,根据所述当前认知层特征和所述自动驾驶控制参数,生成驾驶行为解释。Step S30, generating a driving behavior explanation based on the current cognitive layer characteristics and the automatic driving control parameters.
在本实施例中,需要说明的是,所述自动驾驶控制参数可以包括行驶轨迹、车辆控制参数等,其中,行驶轨迹是指自动驾驶车辆在当前时刻之后一段时间的轨迹,可以控制自动驾驶车辆基于所述行驶轨迹行驶,所述车辆控制参数包括速度、油门踏板开度、方向盘角度等。In this embodiment, it should be noted that the autonomous driving control parameters may include driving trajectory, vehicle control parameters, etc., wherein the driving trajectory refers to the trajectory of the autonomous driving vehicle a period of time after the current moment, and the autonomous driving vehicle can be controlled to drive based on the driving trajectory. The vehicle control parameters include speed, accelerator pedal opening, steering wheel angle, etc.
作为一种示例,所述步骤S30包括:从所述当前认知层特征中可以解码出影响决策层输出自动驾驶控制参数的影响因素的信息,将自动驾驶控制参数和自动驾驶控制参数对应的影响因素的信息进行融合,即可生成驾驶行为解释。As an example, step S30 includes: decoding information about factors affecting the autonomous driving control parameters output by the decision-making layer from the current cognitive layer features, fusing the autonomous driving control parameters and the information about the factors corresponding to the autonomous driving control parameters, and generating a driving behavior explanation.
在一种可实施的方式中,所述根据所述当前认知层特征和所述自动驾驶控制参数,生成驾驶行为解释的步骤之后,还可以包括:输出所述行为解释信息。也即,将所述驾驶行为解释通过图像、声音、文字等方式进行输出,以供用户了解驾驶行为产生的原因。例如,自动驾驶车辆要在红灯路口停车,自动驾驶控制参数可以为停车,影响因素的信息可以为前方路口为红灯以及停止线的位置信息,将二者聚合后可以生成驾驶行为解释“因为前方红灯,所以车辆在停止线前停下来”,可以通过扬声器输出“因为前方红灯,所以车辆在停止线前停下来”的语音播报。In an operative manner, after the step of generating the driving behavior explanation according to the current cognitive layer characteristics and the automatic driving control parameters, the step may further include: outputting the behavior explanation information. That is, the driving behavior explanation is outputted in the form of images, sounds, texts, etc., so that the user can understand the reasons for the driving behavior. For example, if an automatic driving vehicle needs to stop at a red light intersection, the automatic driving control parameter may be to stop, and the information of the influencing factors may be the location information of the red light at the intersection ahead and the stop line. After aggregating the two, a driving behavior explanation "Because the light ahead is red, the vehicle stops before the stop line" can be generated, and a voice broadcast of "Because the light ahead is red, the vehicle stops before the stop line" can be outputted through a speaker.
在一种可实施的方式中,可以在自动驾驶车辆执行所述自动驾驶控制参数之前输出所述驾驶行为解释,这样,可以使得用户预知车辆即将要作出的驾驶行为,就不会因为自动驾驶车辆突然产生的驾驶行为而恐慌。In one feasible manner, the driving behavior explanation may be output before the autonomous driving vehicle executes the autonomous driving control parameters. This allows the user to predict the driving behavior that the vehicle is about to perform and will not panic due to the sudden driving behavior of the autonomous driving vehicle.
可选地,所述自动驾驶控制参数包括多个自动驾驶控制子参数;Optionally, the autonomous driving control parameter includes a plurality of autonomous driving control sub-parameters;
所述根据所述当前认知层特征和所述自动驾驶控制参数,生成驾驶行为解释的步骤包括:The step of generating a driving behavior explanation according to the current cognitive layer characteristics and the automatic driving control parameters comprises:
步骤S31,根据所述当前认知层特征,确定各所述自动驾驶控制子参数各自对应的目标决策特征;Step S31, determining the target decision feature corresponding to each of the autonomous driving control sub-parameters according to the current cognitive layer feature;
步骤S32,将各所述自动驾驶控制子参数与各自对应的目标决策特征分别进行聚合,得到驾驶行为分步解释;Step S32, aggregating each of the autonomous driving control sub-parameters with the corresponding target decision features to obtain a step-by-step explanation of the driving behavior;
步骤S33,将各所述驾驶行为分步解释进行聚合,得到驾驶行为解释。Step S33, aggregating the step-by-step explanations of the driving behavior to obtain a driving behavior explanation.
在本实施例中,需要说明的是,驾驶行为可能是一个步骤即可实现的,也有可能需要多个步骤才能实现,对于需要多个步骤实现的驾驶行为,可以分别确定每个步骤对应的自动驾驶控制子参数,进而按照时间顺序依次排列成,形成自动驾驶控制参数序列。例如,对于加速变道的驾驶行为,所述自动驾驶控制参数序列可以为{τ1(开启右转向灯),τ2(加速),τ3(向右变道)}其中,τ表示自动驾驶控制子参数。对于需要多个步骤实现的驾驶行为,每个步骤均有其执行的原因和关键的影响因素,因此可以分步对每个步骤进行更准确的解释。In this embodiment, it should be noted that the driving behavior may be achieved in one step or may require multiple steps to achieve. For driving behaviors that require multiple steps to achieve, the automatic driving control sub-parameters corresponding to each step can be determined respectively, and then arranged in chronological order to form an automatic driving control parameter sequence. For example, for the driving behavior of accelerating lane change, the automatic driving control parameter sequence can be {τ 1 (turn on the right turn signal), τ 2 (acceleration), τ 3 (change lane to the right)}, where τ represents the automatic driving control sub-parameter. For driving behaviors that require multiple steps to achieve, each step has its execution reasons and key influencing factors, so each step can be explained more accurately step by step.
作为一种示例,所述步骤S31-S33包括:从所述当前认知层特征中可以解码出影响决策层输出各所述自动驾驶控制子参数的影响因素的信息,将每个所述自动驾驶控制子参数对应的影响因素的信息拼接成目标决策特征,将各所述自动驾驶控制子参数与各自对应的目标决策特征进行聚合,生成驾驶行为分步解释,进而将各所述驾驶行为分步解释进一步聚合成驾驶行为解释。As an example, steps S31-S33 include: decoding from the current cognitive layer features information on influencing factors that affect the output of each of the autonomous driving control sub-parameters at the decision layer, splicing the information on the influencing factors corresponding to each of the autonomous driving control sub-parameters into a target decision feature, aggregating each of the autonomous driving control sub-parameters with their respective corresponding target decision features, generating a step-by-step explanation of driving behavior, and further aggregating each of the step-by-step explanations of driving behavior into a driving behavior explanation.
在一种可实施的方式中,所述目标决策特征可以包括浅层目标决策特征和深层目标决策特征,所述根据所述当前认知层特征,确定各所述自动驾驶控制子参数各自对应的目标决策特征的步骤可以包括:可以通过分类模型或采用注意力机制等方式,根据所述当前认知层特征确定各所述自动驾驶控制子参数各自对应的驾驶行为分步解释类型,其中,所述驾驶行为分步解释类型包括浅层解释和深层解释;在确定驾驶行为分步解释类型为浅层解释的情况下,根据所述当前认知层特征,确定各所述自动驾驶控制子参数各自对应的浅层目标决策特征;在确定驾驶行为分步解释类型为深层解释的情况下,根据所述当前认知层特征,确定各所述自动驾驶控制子参数各自对应的深层目标决策特征。其中,浅层解释是指对驾驶行为进行空间上的解释,深层解释是指对驾驶行为进行时空上的解释,每个驾驶行为对应的一系列步骤,对于驾驶行为解释的重要程度是不同的,例如,加速变道的驾驶行为中,由于变道通常都是先减速再变道,因此在加速变道时,用户对于“加速”步骤会产生较大疑惑;且,对于自动驾驶车辆在不同场景遇到不同的实际情况,用户对驾驶行为解释的需求也是不同的,例如,同样是变道,减速变道的驾驶行为符合常规认知,因此不容易产生疑惑,进行浅层解释即可,而加速变道的驾驶行为,通常是在特殊场景中才会出现,因此可能需要对前因后果进行深层解释。In one implementable manner, the target decision feature may include a shallow target decision feature and a deep target decision feature, and the step of determining the target decision feature corresponding to each of the autonomous driving control sub-parameters according to the current cognitive layer feature may include: determining the driving behavior step-by-step explanation type corresponding to each of the autonomous driving control sub-parameters according to the current cognitive layer feature through a classification model or by adopting an attention mechanism, etc., wherein the driving behavior step-by-step explanation type includes a shallow explanation and a deep explanation; when it is determined that the driving behavior step-by-step explanation type is a shallow explanation, determining the shallow target decision feature corresponding to each of the autonomous driving control sub-parameters according to the current cognitive layer feature; when it is determined that the driving behavior step-by-step explanation type is a deep explanation, determining the deep target decision feature corresponding to each of the autonomous driving control sub-parameters according to the current cognitive layer feature. Among them, shallow explanation refers to the spatial explanation of driving behavior, and deep explanation refers to the spatiotemporal explanation of driving behavior. The series of steps corresponding to each driving behavior have different importance for the explanation of driving behavior. For example, in the driving behavior of accelerating and changing lanes, since lane changing usually requires deceleration before changing lanes, when accelerating and changing lanes, users will have great doubts about the "acceleration" step; and, for autonomous driving vehicles encountering different actual situations in different scenarios, users' needs for driving behavior explanations are also different. For example, for the same lane change, the driving behavior of decelerating and changing lanes is in line with conventional cognition, so it is not easy to cause doubts, and a shallow explanation is sufficient. The driving behavior of accelerating and changing lanes usually occurs in special scenarios, so a deep explanation of the cause and effect may be required.
可选地,所述将各所述驾驶行为分步解释进行聚合,得到驾驶行为解释的步骤包括:Optionally, the step of aggregating the step-by-step explanations of the driving behaviors to obtain the driving behavior explanation comprises:
步骤S331,获取各所述自动驾驶控制子参数各自对应的分步解释权重;Step S331, obtaining the step-by-step explanation weight corresponding to each of the autonomous driving control sub-parameters;
步骤S332,基于各所述分步解释权重对各所述驾驶行为分步解释进行加权聚合,得到驾驶行为解释。Step S332: performing weighted aggregation on each of the step-by-step explanations of the driving behavior based on the weight of each of the step-by-step explanations to obtain a driving behavior explanation.
在本实施例中,需要说明的是,用户对于不同步骤的解释的实际需求是不同的,例如,对于加速变道的驾驶行为对应的自动驾驶控制参数序列{τ1(开启右转向灯),τ2(加速),τ3(向右变道)}中,是否开了转向灯用户的感知不明显,而向右变道的原因通常都比较明显,用户抬头看一眼可能就能了解,而加速这个操作在变道过程中是不一定会发生的,为了安全起见,通常是先减速开启转向灯提醒后车,再变道,变道之后再加速,因此在变道之前先加速的情况很容易引起用户警觉,因此,对于加速变道的驾驶行为解释,对加速的操作进行解释的重要性更高,因此可以通过分配权重的方式,使得驾驶行为解释有轻重缓急之分,还可以根据权重确定驾驶行为分步解释的顺序、描述的详细程度等,其中,权重的大小和先后顺序正相关,权重的大小和描述的详细程度也是正相关,以使得权重更高的驾驶行为分布解释可以更早且更详尽地进行输出,而权重较低的驾驶行为分布解释则可以较晚地简单地进行输出,这样既可以保证对较为重要步骤进行详尽且及时的解释,又可以避免对所有步骤进行详细的解释导致用户觉得啰嗦、吵闹而产生反感的情绪。In this embodiment, it should be noted that the actual needs of the user for the explanation of different steps are different. For example, for the driving behavior of accelerating lane change, the automatic driving control parameter sequence {τ 1 (turn on the right turn signal), τ 2 (accelerate), τ 3 (accelerate), τ 4 (accelerate), τ 5 (accelerate), τ 6 (accelerate), τ 7 (accelerate), τ 8 (accelerate), τ 9 (accelerate), In the example (changing lanes to the right), the user's perception of whether the turn signal is on is not obvious, but the reason for changing lanes to the right is usually obvious, and the user may understand it by looking up. The operation of acceleration may not necessarily occur during the lane change process. For safety reasons, it is usually necessary to slow down and turn on the turn signal to remind the following vehicle first, then change lanes, and then accelerate after changing lanes. Therefore, the situation of accelerating before changing lanes is easy to arouse the user's vigilance. Therefore, for the explanation of the driving behavior of accelerating lane change, the explanation of the acceleration operation is more important. Therefore, by allocating weights, the explanation of the driving behavior can be divided into priorities, and the order of the step-by-step explanation of the driving behavior, the level of detail of the description, etc. can also be determined according to the weights. Among them, the size of the weight is positively correlated with the order of precedence, and the size of the weight is also positively correlated with the level of detail of the description, so that the distribution explanation of the driving behavior with a higher weight can be output earlier and more detailed, while the distribution explanation of the driving behavior with a lower weight can be simply output later. In this way, it can ensure that the more important steps are explained in detail and in a timely manner, and it can also avoid the detailed explanation of all steps that makes the user feel verbose and noisy and produce disgusting emotions.
所述分步解释权重是指对自动驾驶控制参数序列中每个自动驾驶控制子参数进行解释,相对于对整个自动驾驶控制参数序列对应的驾驶行为进行解释的重要程度。可以预先根据测试结果、实际需要等通过人工或对应的模型进行设置,也可以采用注意力机制基于各所述驾驶行为分步解释和上一时间步确定的注意力分布特征进行确定,本实施例对此不加以限制。The step-by-step explanation weight refers to the importance of explaining each autonomous driving control sub-parameter in the autonomous driving control parameter sequence relative to explaining the driving behavior corresponding to the entire autonomous driving control parameter sequence. It can be set in advance by manual or corresponding models according to test results, actual needs, etc., or it can be determined by using an attention mechanism based on the step-by-step explanation of each driving behavior and the attention distribution characteristics determined in the previous time step. This embodiment does not limit this.
作为一种示例,所述步骤S331-S332包括:获取当前确定的所述自动驾驶控制参数中各个自动驾驶控制子参数各自对应的分步解释权重,基于各所述分步解释权重对各所述驾驶行为分步解释进行加权聚合,得到驾驶行为解释。As an example, the steps S331-S332 include: obtaining the step-by-step explanation weights corresponding to each autonomous driving control sub-parameter in the currently determined autonomous driving control parameters, and performing weighted aggregation on each step-by-step explanation of the driving behavior based on each step-by-step explanation weight to obtain a driving behavior explanation.
在一种可实施的方式中,所述采用注意力机制,基于各所述驾驶行为分步解释以及所述上一时间步的注意力分布特征,确定各所述自动驾驶控制子参数各自对应的分步解释权重的步骤可以表示为:In an practicable manner, the step of using the attention mechanism to determine the step-by-step explanation weight corresponding to each of the autonomous driving control sub-parameters based on the step-by-step explanation of each driving behavior and the attention distribution characteristics of the previous time step can be expressed as:
其中,J是指驾驶行为解释,βt是指t行为时刻对应的分步解释权重,yt是指t行为时刻对应的驾驶行为分步解释,t∈[t0,T]。Where J refers to the driving behavior explanation, βt refers to the step-by-step explanation weight corresponding to the behavior time t, yt refers to the step-by-step explanation of the driving behavior corresponding to the behavior time t, t∈[t 0 ,T].
可选地,所述获取各所述自动驾驶控制子参数各自对应的分步解释权重的步骤包括:Optionally, the step of obtaining the step-by-step interpretation weight corresponding to each of the autonomous driving control sub-parameters includes:
步骤S3311,获取上一时间步的注意力分布特征;Step S3311, obtaining the attention distribution feature of the previous time step;
步骤S3312,采用注意力机制,基于各所述驾驶行为分步解释以及所述上一时间步的注意力分布特征,确定各所述自动驾驶控制子参数各自对应的分步解释权重。Step S3312, using the attention mechanism, based on the step-by-step interpretation of each driving behavior and the attention distribution characteristics of the previous time step, determines the step-by-step interpretation weight corresponding to each of the autonomous driving control sub-parameters.
作为一种示例,所述步骤S3311-S3312包括:获取上一时间步的注意力分布特征,进而采用注意力机制,对各所述驾驶行为分步解释和上一时间步的注意力分布特征进行特征融合、特征编码、特征推理等,确定当前时间步的分步解释权重,将当前时间步的分步解释权重确定为当前确定的所述自动驾驶控制参数中的各所述自动驾驶控制子参数各自对应的分步解释权重。As an example, steps S3311-S3312 include: obtaining the attention distribution characteristics of the previous time step, and then using the attention mechanism to perform feature fusion, feature encoding, feature reasoning, etc. on the step-by-step interpretation of each driving behavior and the attention distribution characteristics of the previous time step, determine the step-by-step interpretation weight of the current time step, and determine the step-by-step interpretation weight of the current time step as the step-by-step interpretation weight corresponding to each of the autonomous driving control sub-parameters in the currently determined autonomous driving control parameters.
在一种可实施的方式中,参照图2,所述自动驾驶控制装置包括座舱域和智驾域人机交互接口、认知决策模型和车辆驾驶行为体现模块,其中,所述座舱域和智驾域人机交互接口用于接收座舱域和智驾域获取用户需求信息和环境认知信息,所述认知决策模型用于确定自动驾驶控制参数,所述车辆驾驶行为体现模块用于控制自动驾驶车辆基于所述自动驾驶控制参数作出相应的驾驶行为,在作出驾驶行为之后,又可以通过所述座舱域和智驾域人机交互接口获得用户对于驾驶行为的反馈信息,如此循环,可以使得自动驾驶车辆作出适应于用户需求的驾驶行为。In one implementable manner, referring to Figure 2, the autonomous driving control device includes a cockpit domain and an intelligent driving domain human-computer interaction interface, a cognitive decision-making model and a vehicle driving behavior embodiment module, wherein the cockpit domain and the intelligent driving domain human-computer interaction interface is used to receive user demand information and environmental cognition information obtained by the cockpit domain and the intelligent driving domain, the cognitive decision-making model is used to determine the autonomous driving control parameters, and the vehicle driving behavior embodiment module is used to control the autonomous driving vehicle to perform corresponding driving behaviors based on the autonomous driving control parameters. After performing the driving behavior, the user's feedback information on the driving behavior can be obtained through the cockpit domain and the intelligent driving domain human-computer interaction interface. Such a cycle can enable the autonomous driving vehicle to perform driving behaviors that adapt to user needs.
在本实施例中,所述自动驾驶控制方法采用认知决策模型,所述认知决策模型包括认知层和决策层。首先,通过获取车内外监测数据,通过所述认知层从所述车内外监测数据中提取当前认知层特征,实现了对自动驾驶车辆当前情况的认知;进而通过将所述环境认知特征和所述用户需求特征输入所述决策层,确定自动驾驶控制参数,实现了在认知到当前情况的基础上,准确作出自动驾驶决策,确定自动驾驶控制参数;进而通过根据所述当前认知层特征和所述自动驾驶控制参数,生成驾驶行为解释,实现对自动驾驶决策进行解释的目的。由于自动驾驶控制参数是基于认知层特征进行决策确定的,因此自动驾驶控制参数与认知层特征之间存在一定的因果关系,而用户之所以无法理解自动驾驶车辆的驾驶行为,正是因为用户在感受到自动驾驶车辆突发性的驾驶行为之前,注意力可能与驾驶车辆的注意力集中在不同的事务上,因此在感受到自动驾驶车辆突发性的驾驶行为之时,感受到的是自动驾驶车辆最终作出的决策,可能看不到影响决策的因素,从而无法推理出其作出决策的原因,因此会产生疑惑、恐慌、不安等负面情绪。因此,本申请基于认知层特征对自动驾驶控制参数进行解释,生成驾驶行为解释,可以主动与用户进行交互,为用户解释驾驶行为产生的原因,消除用户对驾驶行为的疑惑。因此克服了端到端感知决策架构的可解释性较差,因此在控制车辆执行突发性行为时,容易导致乘坐自动驾驶车辆的用户无法理解自动驾驶车辆的驾驶行为而产生疑惑、恐慌、不安等负面情绪,用户体验感较差的技术缺陷,提高了端到端感知决策架构的可解释性,减少了用户因未知而产生的疑惑、恐慌、不安等负面情绪,提高了用户体验感。In this embodiment, the automatic driving control method adopts a cognitive decision model, and the cognitive decision model includes a cognitive layer and a decision layer. First, by acquiring the monitoring data inside and outside the vehicle, the cognitive layer extracts the current cognitive layer features from the monitoring data inside and outside the vehicle, so as to realize the recognition of the current situation of the automatic driving vehicle; then, by inputting the environmental cognitive features and the user demand features into the decision layer, the automatic driving control parameters are determined, so as to realize the accurate automatic driving decision and determine the automatic driving control parameters based on the recognition of the current situation; then, by generating the driving behavior explanation according to the current cognitive layer features and the automatic driving control parameters, the purpose of explaining the automatic driving decision is realized. Since the automatic driving control parameters are determined based on the cognitive layer features, there is a certain causal relationship between the automatic driving control parameters and the cognitive layer features. The reason why the user cannot understand the driving behavior of the automatic driving vehicle is that before the user feels the sudden driving behavior of the automatic driving vehicle, the user's attention may be focused on different matters from the driving vehicle. Therefore, when the user feels the sudden driving behavior of the automatic driving vehicle, the user feels the final decision made by the automatic driving vehicle, and may not see the factors affecting the decision, so as to be unable to infer the reason for its decision, so as to generate negative emotions such as doubt, panic, and uneasiness. Therefore, this application interprets the autonomous driving control parameters based on the cognitive layer characteristics, generates a driving behavior explanation, can actively interact with the user, explain the reasons for the driving behavior to the user, and eliminate the user's doubts about the driving behavior. Therefore, it overcomes the poor interpretability of the end-to-end perception decision architecture. Therefore, when controlling the vehicle to perform sudden behaviors, it is easy to cause users riding in the autonomous driving vehicle to be unable to understand the driving behavior of the autonomous driving vehicle and generate negative emotions such as doubt, panic, and uneasiness, and the technical defects of poor user experience. It improves the interpretability of the end-to-end perception decision architecture, reduces the user's negative emotions such as doubt, panic, and uneasiness caused by the unknown, and improves the user experience.
实施例二Embodiment 2
进一步地,在本申请的第二实施例中,与上述实施例相同或相似的内容,可以参考上文介绍,后续不再赘述。在此基础上,参照图3,所述当前认知层特征包括车外环境特征;所述目标决策特征包括浅层目标决策特征;Further, in the second embodiment of the present application, the same or similar contents as those in the above embodiment can be referred to the above description, and will not be described in detail later. On this basis, referring to FIG. 3 , the current cognitive layer features include the features of the vehicle's external environment; the target decision features include the shallow target decision features;
所述根据所述当前认知层特征,确定各所述自动驾驶控制子参数各自对应的目标决策特征的步骤包括:The step of determining the target decision feature corresponding to each of the autonomous driving control sub-parameters according to the current cognitive layer feature comprises:
步骤B10,根据所述当前认知层特征,从各所述自动驾驶控制子参数中确定至少一个浅层待解释子参数,以及各所述浅层待解释子参数各自对应的空间注意力权重分布;Step B10, determining at least one shallow-layer sub-parameter to be explained from each of the autonomous driving control sub-parameters according to the current cognitive layer characteristics, and the spatial attention weight distribution corresponding to each of the shallow-layer sub-parameters to be explained;
在本实施例中,需要说明的是,自动驾驶行为解释并非是越多越好,在用户无法理解自动驾驶车辆的驾驶行为的情况下,对驾驶行为进行解释,可以有效环境用户的负面情绪,但若对于容易理解的驾驶行为,仍进行详尽的解释,一方面,越详尽的解释会需要进行越多的分析,消耗越多的算力,另一方面,对于很容易理解的驾驶行为仍进行很详细的解释,反而容易导致用户觉得啰嗦、吵闹、分不清重点而产生反感的情绪。In this embodiment, it should be noted that the more explanations of autonomous driving behaviors are provided, the better. When users cannot understand the driving behavior of the autonomous driving vehicle, explaining the driving behavior can effectively alleviate the user's negative emotions. However, if a detailed explanation is still given for driving behaviors that are easy to understand, on the one hand, the more detailed the explanation, the more analysis will be required and the more computing power will be consumed. On the other hand, a very detailed explanation of driving behaviors that are easy to understand may easily cause users to feel that the explanation is verbose, noisy, and unable to distinguish the key points, which may lead to negative emotions.
所述空间注意力权重分布用于表征自动驾驶车辆在当前场景当前步骤实现安全驾驶所需的注意力分布情况,具有高注意力权重的位置对于任务是显著的。The spatial attention weight distribution is used to characterize the attention distribution required for the autonomous driving vehicle to achieve safe driving in the current scene and the current step. Positions with high attention weights are significant for the task.
作为一种示例,所述步骤B10包括:可以通过分类模型或采用注意力机制等方式,根据所述当前认知层特征从各所述自动驾驶控制子参数中确定至少一个浅层待解释子参数,同步地,可以通过激活函数基于所述当前认知层特征输出各所述浅层待解释子参数各自对应的空间注意力权重分布。As an example, the step B10 includes: determining at least one shallow sub-parameter to be interpreted from each of the autonomous driving control sub-parameters according to the current cognitive layer characteristics through a classification model or an attention mechanism, and simultaneously, outputting the spatial attention weight distribution corresponding to each of the shallow sub-parameters to be interpreted based on the current cognitive layer characteristics through an activation function.
步骤B20,根据所述空间注意力权重分布和所述车外环境特征确定各所述浅层待解释子参数各自对应的浅层目标决策特征。Step B20, determining the shallow target decision features corresponding to each of the shallow sub-parameters to be interpreted according to the spatial attention weight distribution and the external vehicle environment characteristics.
作为一种示例,所述步骤B20包括:融合各所述浅层待解释子参数各自对应的所述空间注意力权重分布和各自对应的所述车外环境特征,即可得到各所述浅层待解释子参数各自对应的浅层目标决策特征。As an example, step B20 includes: fusing the spatial attention weight distribution corresponding to each of the shallow sub-parameters to be interpreted and the corresponding external vehicle environment characteristics, so as to obtain the shallow target decision features corresponding to each of the shallow sub-parameters to be interpreted.
在一种可实施方式中,所述根据所述空间注意力权重分布和所述车外环境特征确定各所述浅层待解释子参数各自对应的浅层目标决策特征的步骤可以表示为: In one possible implementation, the step of determining the shallow target decision features corresponding to each of the shallow to-be-interpreted sub-parameters according to the spatial attention weight distribution and the vehicle exterior environment features can be expressed as:
其中,s表示浅层,是指t行为时刻的浅层待解释子参数,是t行为时刻第i个自动驾驶控制子参数的空间注意力权重分布图,并满足xt,i是t行为时刻第i个自动驾驶控制子参数对应的车外环境特征,i={1,2,…l},l是指自动驾驶控制子参数的数量。Among them, s represents the shallow layer, refers to the shallow sub-parameter to be explained at the time of behavior t, is the spatial attention weight distribution diagram of the ith autonomous driving control sub-parameter at behavior time t, and satisfies x t,i is the external environment characteristic corresponding to the i-th autonomous driving control sub-parameter at behavior time t, i = {1, 2, … l}, l refers to the number of autonomous driving control sub-parameters.
可选地,所述获取车内外监测数据,通过所述认知层从所述车内外监测数据中提取当前认知层特征的步骤之前,还包括:Optionally, before the step of acquiring the vehicle interior and exterior monitoring data and extracting the current cognitive layer features from the vehicle interior and exterior monitoring data through the cognitive layer, the step further includes:
步骤C10,获取训练样本车内外监测数据和自动驾驶行为标签,通过所述认知层从所述训练样本车内外监测数据中提取训练样本当前认知层特征;Step C10, obtaining the in-vehicle and out-vehicle monitoring data and the autonomous driving behavior label of the training sample, and extracting the current cognitive layer features of the training sample from the in-vehicle and out-vehicle monitoring data of the training sample through the cognitive layer;
步骤C20,将所述训练样本当前认知层特征输入所述决策层,确定多个训练样本自动驾驶控制子参数;Step C20, inputting the current cognitive layer features of the training samples into the decision layer, and determining multiple automatic driving control sub-parameters of the training samples;
步骤C30,根据所述训练样本当前认知层特征,从各所述训练样本自动驾驶控制子参数中确定至少一个训练样本浅层待解释子参数,以及各所述训练样本浅层待解释子参数各自对应的训练样本空间注意力权重分布;Step C30, determining at least one shallow sub-parameter to be explained of the training sample from each of the autonomous driving control sub-parameters of the training sample according to the current cognitive layer characteristics of the training sample, and the spatial attention weight distribution of the training samples corresponding to each of the shallow sub-parameters to be explained of the training sample;
步骤C40,基于各所述训练样本自动驾驶控制子参数、各所述自动驾驶行为标签以及各所述训练样本空间注意力权重分布确定浅层注意力损失,基于所述浅层注意力损失对所述认知决策模型进行迭代优化。Step C40, determining the shallow attention loss based on the autonomous driving control sub-parameters of each training sample, the autonomous driving behavior labels, and the spatial attention weight distribution of each training sample, and iteratively optimizing the cognitive decision-making model based on the shallow attention loss.
在本实施例中,需要说明的是,在认知决策模型应用之前,还需要对认知决策模型进行模型训练,在模型训练的每一轮迭代优化的过程中,都可以基于计算得到的模型损失对认知决策模型进行更新,直至认知决策模型收敛。认知决策模型的训练过程可以在最终应用的设备上进行,考虑到模型训练算力的需求,也可以在其他设备上训练完成之后,将训练好的认知决策模型部署到应用设备上进行应用。例如,可以在服务器上对认知决策模型进行训练,进而将训练好的认知决策模型部署于自动驾驶车辆上进行应用。模型训练过程中基于训练样本进行认知决策的方式与模型应用过程相近,与上述实施例相同或相似的内容,可以参考上文介绍,后续不再赘述。在所述目标决策特征包括浅层目标决策特征的情况下,由于自动驾驶行为预测与空间注意力权重分布息息相关,因此可以利用训练样本空间注意力权重分布的检测准确性对自动驾驶行为预测损失进行修正,降低单一的计算驾驶行为预测损失的误差。In this embodiment, it should be noted that before the cognitive decision model is applied, the cognitive decision model needs to be trained. In each round of iterative optimization of the model training, the cognitive decision model can be updated based on the calculated model loss until the cognitive decision model converges. The training process of the cognitive decision model can be carried out on the device of the final application. Considering the demand for model training computing power, the trained cognitive decision model can also be deployed to the application device for application after the training is completed on other devices. For example, the cognitive decision model can be trained on the server, and then the trained cognitive decision model can be deployed on the autonomous driving vehicle for application. The way of making cognitive decisions based on training samples during the model training process is similar to the model application process. For the same or similar content as the above embodiment, please refer to the above introduction, and no further description will be given later. In the case where the target decision feature includes a shallow target decision feature, since the autonomous driving behavior prediction is closely related to the spatial attention weight distribution, the detection accuracy of the spatial attention weight distribution of the training sample can be used to correct the autonomous driving behavior prediction loss, thereby reducing the error of the single calculation of the driving behavior prediction loss.
作为一种示例,所述步骤C10-C40包括:获取训练样本车内外监测数据和自动驾驶行为标签,通过所述认知层从所述训练样本车内外监测数据中提取训练样本当前认知层特征;将所述训练样本当前认知层特征输入所述决策层,确定多个训练样本自动驾驶控制子参数;根据所述训练样本当前认知层特征,从各所述训练样本自动驾驶控制子参数中确定至少一个训练样本浅层待解释子参数,以及各所述训练样本浅层待解释子参数各自对应的训练样本空间注意力权重分布;基于各所述训练样本自动驾驶控制子参数与各自对应的所述自动驾驶行为标签之间的差异,确定至少一个行为时刻对应的初始预测损失,并基于各所述行为时刻各自对应的训练样本空间注意力权重分布确定空间注意力损失,基于所述空间注意力损失对各自对应的初始预测损失进行修正,修正后得到浅层注意力损失;进而,基于所述浅层注意力损失判断认知决策模型是否收敛,在确定认知决策模型收敛的情况下,可以确定模型训练完成,得到训练完成的认知决策模型;在确定认知决策模型未收敛的情况下,可以根据梯度下降法基于所述浅层注意力损失对认知决策模型进行一轮更新,并返回执行所述获取训练样本车内外监测数据和自动驾驶行为标签的步骤,进行下一轮的训练。As an example, the steps C10-C40 include: obtaining in-vehicle and out-vehicle monitoring data and autonomous driving behavior labels of the training samples, extracting current cognitive layer features of the training samples from the in-vehicle and out-vehicle monitoring data of the training samples through the cognitive layer; inputting the current cognitive layer features of the training samples into the decision layer to determine multiple autonomous driving control sub-parameters of the training samples; determining at least one shallow sub-parameter to be explained of the training samples from each of the autonomous driving control sub-parameters of the training samples according to the current cognitive layer features of the training samples, and the spatial attention weight distribution of the training samples corresponding to each of the shallow sub-parameters to be explained of the training samples; determining at least one based on the difference between each of the autonomous driving control sub-parameters of the training samples and the corresponding autonomous driving behavior labels; The initial prediction loss corresponding to each behavior moment is determined, and the spatial attention loss is determined based on the spatial attention weight distribution of the training samples corresponding to each of the behavior moments. The initial prediction loss corresponding to each behavior moment is corrected based on the spatial attention loss to obtain the shallow attention loss after correction. Then, whether the cognitive decision model has converged is judged based on the shallow attention loss. When it is determined that the cognitive decision model has converged, it can be determined that the model training is completed to obtain a trained cognitive decision model. When it is determined that the cognitive decision model has not converged, a round of updating of the cognitive decision model can be performed based on the shallow attention loss according to the gradient descent method, and the step of obtaining the in-vehicle and out-of-vehicle monitoring data and the autonomous driving behavior labels of the training samples is returned to perform the next round of training.
在一种可实施的方式中,所述基于所述空间注意力损失对各自对应的初始预测损失进行修正的方式可以为,基于各所述行为时刻各自对应的训练样本空间注意力权重分布计算注意力权重熵值,将所述注意力权重熵值确定为空间注意力损失,将所述空间注意力损失与各所述初始预测损失进行聚合,以实现对各所述初始预测损失的修正。需要说明的是,注意力权重分布用于表示输入图像中每个元素的重要性,所述注意力权重熵值是指注意力权重分布的熵值,用来度量模型对于输入图像中各个元素的关注程度的不确定性,如果注意力权重分布集中分配给一个或少数几个元素,则计算确定的注意力权重熵值会比较低,表示模型对输入图像的关注是高度确定的,相反,如果注意力权重分布比较均匀,则计算确定的注意力权重熵值会比较高,表示模型对输入图像的关注是不确定的。In an operative manner, the method for correcting the respective corresponding initial prediction losses based on the spatial attention loss can be, based on the spatial attention weight distribution of the training samples corresponding to each of the behavior moments, calculating the attention weight entropy value, determining the attention weight entropy value as the spatial attention loss, and aggregating the spatial attention loss with each of the initial prediction losses to achieve correction of each of the initial prediction losses. It should be noted that the attention weight distribution is used to represent the importance of each element in the input image, and the attention weight entropy value refers to the entropy value of the attention weight distribution, which is used to measure the uncertainty of the model's attention to each element in the input image. If the attention weight distribution is concentrated on one or a few elements, the calculated attention weight entropy value will be relatively low, indicating that the model's attention to the input image is highly certain. On the contrary, if the attention weight distribution is relatively uniform, the calculated attention weight entropy value will be relatively high, indicating that the model's attention to the input image is uncertain.
可选地,所述基于各所述训练样本自动驾驶控制子参数、各所述自动驾驶行为标签以及各所述训练样本空间注意力权重分布确定浅层注意力损失的步骤包括:Optionally, the step of determining the shallow attention loss based on each of the training sample autonomous driving control sub-parameters, each of the autonomous driving behavior labels, and each of the training sample spatial attention weight distributions includes:
将各所述训练样本自动驾驶控制子参数、各所述自动驾驶行为标签以及各所述训练样本空间注意力权重分布输入浅层注意力损失函数,确定浅层注意力损失,其中,所述浅层注意力损失函数为:Input each of the training sample autonomous driving control sub-parameters, each of the autonomous driving behavior labels, and each of the training sample spatial attention weight distributions into a shallow attention loss function to determine a shallow attention loss, wherein the shallow attention loss function is:
其中,为浅层注意力损失,τt为t行为时刻的自动驾驶行为标签,为t行为时刻的训练样本自动驾驶控制子参数,H为熵函数,用于计算训练样本空间注意力权重分布的熵,λs为预设第一超参数,为t行为时刻的训练样本空间注意力权重分布,t∈[t0,T]。in, is the shallow attention loss, τt is the autonomous driving behavior label at behavior time t, is the sub-parameter of the autonomous driving control of the training sample at the behavior time t, H is the entropy function used to calculate the entropy of the spatial attention weight distribution of the training sample, λ s is the preset first hyperparameter, is the spatial attention weight distribution of the training sample at behavior time t, t∈[t 0 ,T].
在本实施例中,通过浅层解释机制,可以对用户容易理解驾驶行为进行空间层面的浅层解释,一方面,空间层面的解释仅需要利用当前时刻的认知层特征,而无需进行时间维度的分析,一方面,可以降低算力消耗,提高生成驾驶行为解释的效率,另一方面,可以简化驾驶行为解释,避免用户觉得啰嗦、吵闹、分不清重点而产生反感的情绪。例如,“在红绿灯路口停车”的驾驶行为,决策的关键影响因素是前方的红灯,空间注意力权重分布中较大权重也会集中在红绿灯的位置,因此进行空间层面的浅层解释“因为前方红灯,所以车辆在停止线前停下来”即可。In this embodiment, through the shallow explanation mechanism, a shallow explanation at the spatial level can be made for driving behaviors that are easy for users to understand. On the one hand, the explanation at the spatial level only needs to use the cognitive layer features of the current moment, without the need for time dimension analysis. On the one hand, it can reduce computing power consumption and improve the efficiency of generating driving behavior explanations. On the other hand, it can simplify the driving behavior explanations to avoid users feeling verbose, noisy, and unable to distinguish the key points and generating negative emotions. For example, for the driving behavior of "stopping at a traffic light intersection", the key influencing factor of the decision is the red light ahead, and the larger weight in the spatial attention weight distribution will also be concentrated on the position of the traffic light. Therefore, a shallow explanation at the spatial level "because the red light ahead, the vehicle stops before the stop line" can be made.
实施例三Embodiment 3
进一步地,在本申请的第三实施例中,与上述实施例相同或相似的内容,可以参考上文介绍,后续不再赘述。在此基础上,参照图4,所述当前认知层特征包括车外环境特征;所述目标决策特征包括浅层目标决策特征;Further, in the third embodiment of the present application, the same or similar contents as those in the above embodiment can be referred to the above description, and will not be described in detail later. On this basis, referring to FIG. 4 , the current cognitive layer features include the features of the vehicle's external environment; the target decision features include the shallow target decision features;
所述根据所述当前认知层特征,确定各所述自动驾驶控制子参数各自对应的目标决策特征的步骤包括:The step of determining the target decision feature corresponding to each of the autonomous driving control sub-parameters according to the current cognitive layer feature comprises:
步骤D10,根据所述当前认知层特征,从各所述自动驾驶控制子参数中确定至少一个深层待解释子参数;Step D10, determining at least one deep-level sub-parameter to be explained from each of the autonomous driving control sub-parameters according to the current cognitive layer characteristics;
在本实施例中,需要说明的是,若对复杂场景下作出的驾驶行为决策仅进行空间层面的浅层解释,用户需要自己去思考和关联其中的时空逻辑,且可能由于缺失部分历史信息,例如突然跑过马路上的猫,用户感受到刹车之后再抬头去看时,猫可能已经跑走了,从而可能无法准确理解驾驶行为而产生疑惑、恐慌、不安等负面情绪,用户体验感较差。In this embodiment, it should be noted that if the driving behavior decisions made in complex scenarios are only explained superficially at the spatial level, the user needs to think and associate the spatiotemporal logic by himself, and some historical information may be missing. For example, if a cat suddenly runs across the road, when the user feels the brakes and looks up, the cat may have run away. As a result, the user may not be able to accurately understand the driving behavior and may have negative emotions such as doubt, panic, and anxiety, resulting in a poor user experience.
作为一种示例,所述步骤D10包括:可以通过分类模型或采用注意力机制等方式,根据所述当前认知层特征从各所述自动驾驶控制子参数中确定至少一个深层待解释子参数。As an example, the step D10 includes: determining at least one deep-level sub-parameter to be explained from each of the autonomous driving control sub-parameters according to the current cognitive layer characteristics through a classification model or an attention mechanism.
步骤D20,获取当前时刻之前提取的历史认知层特征;Step D20, obtaining historical cognitive layer features extracted before the current moment;
步骤D30,采用时空注意力机制,基于所述当前认知层特征和所述历史认知层特征,提取各所述深层待解释子参数各自对应的深层目标决策特征。Step D30, adopting the spatiotemporal attention mechanism, based on the current cognitive layer features and the historical cognitive layer features, extracts the deep target decision features corresponding to each of the deep sub-parameters to be explained.
在本实施例中,需要说明的是,所述历史认知层特征是指当前时刻之前提取到的认知层特征。深层解释是针对于那些通过空间维度无法清楚且直观地进行解释的驾驶行为,进行时间和空间两个维度的解释,所述深层目标决策特征中包含有影响决策的影响因素的空间信息和时间信息,例如可以包含有运动趋势信息、运动轨迹信息等。若要进行时间维度的解释,不论是对过去到现在进行分析,还是对未来进行预测,都需要基于过去到现在产生的历史认知层特征进行趋势分析,才能更准确地确定对驾驶行为产生影响的关键因素,因此,需要获取历史认知层特征,并结合当前认知层特征进行分析。例如,“人车混行场景车辆加速并变道避障”的驾驶行为,因为涉及的影响因素较多,且车辆产生了在变道之前先加速的非常规操作,如果只进行空间维度的浅层解释“前方有人所有加速变道避障”,但因为前方有人的情况下,通常应该减速让行,因此用户仍然会对自动驾驶车辆的驾驶行为产生疑惑,这时就需要进行时间和空间两个维度的解释了,生成的深层解释可以为“因为行人的运动轨迹与本车轨迹将会交叉,有碰撞风险,本车道无法减速避撞,所以加速变道进行避障”,还可以输出较大权重集中在无法避让的行人的位置的空间注意力权重分布图,让用户更直观地看到无法避让的行人,从而更好地理解驾驶行为。In this embodiment, it should be noted that the historical cognitive layer features refer to the cognitive layer features extracted before the current moment. The deep interpretation is to interpret the two dimensions of time and space for those driving behaviors that cannot be clearly and intuitively explained through the spatial dimension. The deep target decision-making features contain spatial information and time information of the factors that affect the decision, such as motion trend information, motion trajectory information, etc. If the time dimension interpretation is to be performed, whether it is to analyze from the past to the present or to predict the future, it is necessary to perform trend analysis based on the historical cognitive layer features generated from the past to the present in order to more accurately determine the key factors that affect driving behavior. Therefore, it is necessary to obtain the historical cognitive layer features and analyze them in combination with the current cognitive layer features. For example, the driving behavior of "accelerating and changing lanes to avoid obstacles in a mixed traffic scene" involves many influencing factors and the vehicle performs an unconventional operation of accelerating before changing lanes. If only a shallow explanation of the spatial dimension is performed, "there is someone ahead, so accelerate and change lanes to avoid obstacles", but because there is someone ahead, you should usually slow down and give way, so users will still be confused about the driving behavior of the autonomous vehicle. At this time, explanations in both time and space dimensions are needed. The generated deep explanation can be "because the movement trajectory of the pedestrian will intersect with the trajectory of this vehicle, there is a risk of collision, and this lane cannot slow down to avoid collision, so accelerate and change lanes to avoid obstacles". It can also output a spatial attention weight distribution map with a large weight concentrated on the position of pedestrians that cannot be avoided, allowing users to more intuitively see the pedestrians that cannot be avoided, and thus better understand driving behavior.
作为一种示例,所述步骤D20包括:可以采用时空注意力机制,对所述当前认知层特征、所述历史认知层特征以及各所述深层待解释子参数进行特征融合、特征编码、特征推理等分析,确定各所述深层待解释子参数各自对应的深层目标决策特征。As an example, step D20 includes: a spatiotemporal attention mechanism can be used to perform feature fusion, feature encoding, feature reasoning and other analyses on the current cognitive layer features, the historical cognitive layer features and each of the deep-layer sub-parameters to be interpreted, so as to determine the deep target decision features corresponding to each of the deep-layer sub-parameters to be interpreted.
可选地,所述基于各所述训练样本自动驾驶控制子参数、各所述自动驾驶行为标签以及各所述训练样本空间注意力权重分布确定浅层注意力损失,基于所述浅层注意力损失对所述认知决策模型进行迭代优化的步骤包括:Optionally, the step of determining a shallow attention loss based on each of the training sample autonomous driving control sub-parameters, each of the autonomous driving behavior labels, and each of the training sample spatial attention weight distributions, and iteratively optimizing the cognitive decision model based on the shallow attention loss includes:
步骤C41,根据所述训练样本当前认知层特征,从各所述训练样本自动驾驶控制子参数中确定至少一个训练样本深层待解释子参数、以及各所述训练样本深层待解释子参数各自对应的训练样本空间注意力权重分布;Step C41, determining at least one training sample deep-layer to-be-explained sub-parameter from each of the training sample autonomous driving control sub-parameters according to the current cognitive layer characteristics of the training sample, and the spatial attention weight distribution of the training samples corresponding to each of the training sample deep-layer to-be-explained sub-parameters;
步骤C42,获取当前时刻之前提取的训练样本历史认知层特征,采用时空注意力机制,基于所述训练样本当前认知层特征和所述训练样本历史认知层特征,确定各所述训练样本深层待解释子参数各自对应的训练样本时空注意力权重分布;Step C42, obtaining the historical cognitive layer features of the training samples extracted before the current moment, using the spatiotemporal attention mechanism, based on the current cognitive layer features of the training samples and the historical cognitive layer features of the training samples, determining the spatiotemporal attention weight distribution of the training samples corresponding to the deep-layer to-be-explained sub-parameters of the training samples;
步骤C43,基于各所述训练样本深层待解释子参数各自对应的训练样本空间注意力权重分布与各自对应的训练样本时空注意力权重分布之间的差异确定深层注意力损失,基于各所述训练样本空间注意力权重分布以及各所述训练样本自动驾驶控制子参数与各自对应的自动驾驶行为标签之间的差异,确定浅层注意力损失;Step C43, determining the deep attention loss based on the difference between the spatial attention weight distribution of the training samples corresponding to the deep sub-parameters to be interpreted of the training samples and the corresponding spatiotemporal attention weight distribution of the training samples, and determining the shallow attention loss based on the difference between the spatial attention weight distribution of the training samples and the autonomous driving control sub-parameters of the training samples and the corresponding autonomous driving behavior labels;
步骤C44,将所述深层注意力损失和所述浅层注意力损失进行聚合,得到自动驾驶行为预测损失,基于所述自动驾驶行为预测损失对所述认知决策模型进行迭代优化。Step C44, aggregating the deep attention loss and the shallow attention loss to obtain the autonomous driving behavior prediction loss, and iteratively optimizing the cognitive decision model based on the autonomous driving behavior prediction loss.
在本实施例中,需要说明的是,在所述目标决策特征包括深层目标决策特征的情况下,由于不论是进行空间维度的分析还是时间空间两个维度的分析,都需要确定影响决策的关键因素,这一关键因素通常是根据注意力权重分布进行确定,因此可以通过对比从空间维度确定的空间注意力权重分布与从时空维度确定的时空注意力权重分布之间的差异,使得时空注意力机制参考显著的关注对象,也即空间维度确定的空间注意力权重分布中权重较高的位置,并在时空维度上对认知决策模型输出的车辆驾驶行为进行时间和空间两个维度的解释。In this embodiment, it should be noted that when the target decision feature includes a deep target decision feature, since it is necessary to determine the key factors affecting the decision regardless of whether the analysis is in the spatial dimension or in the time and space dimensions, this key factor is usually determined based on the attention weight distribution. Therefore, by comparing the difference between the spatial attention weight distribution determined from the spatial dimension and the time and space attention weight distribution determined from the time and space dimension, the time and space attention mechanism can refer to the significant object of attention, that is, the position with a higher weight in the spatial attention weight distribution determined in the spatial dimension, and interpret the vehicle driving behavior output by the cognitive decision model in the time and space dimensions.
作为一种示例,所述步骤C41-C44包括:根据所述训练样本当前认知层特征,从各所述训练样本自动驾驶控制子参数中确定至少一个训练样本深层待解释子参数、以及各所述训练样本深层待解释子参数各自对应的训练样本空间注意力权重分布;获取当前时刻之前提取的训练样本历史认知层特征,采用时空注意力机制,基于所述训练样本当前认知层特征和所述训练样本历史认知层特征,确定各所述训练样本深层待解释子参数各自对应的训练样本时空注意力权重分布;基于各所述训练样本深层待解释子参数各自对应的训练样本空间注意力权重分布与各自对应的训练样本时空注意力权重分布之间的差异确定深层注意力损失,基于各所述训练样本空间注意力权重分布以及各所述训练样本自动驾驶控制子参数与各自对应的自动驾驶行为标签之间的差异,确定浅层注意力损失;进而将所述深层注意力损失和所述浅层注意力损失进行聚合,得到自动驾驶行为预测损失;进而,基于所述自动驾驶行为预测损失判断认知决策模型是否收敛,在确定认知决策模型收敛的情况下,可以确定模型训练完成,得到训练完成的认知决策模型;在确定认知决策模型未收敛的情况下,可以根据梯度下降法基于所述自动驾驶行为预测损失对认知决策模型进行一轮更新,并返回执行所述获取训练样本车内外监测数据和自动驾驶行为标签的步骤,进行下一轮的训练。As an example, the steps C41-C44 include: determining at least one deep sub-parameter to be interpreted of the training samples and the spatial attention weight distribution of the training samples corresponding to each of the deep sub-parameters to be interpreted of the training samples from each of the autonomous driving control sub-parameters of the training samples according to the current cognitive layer characteristics of the training samples; obtaining the historical cognitive layer characteristics of the training samples extracted before the current moment, and using the spatiotemporal attention mechanism to determine the spatiotemporal attention weight distribution of the training samples corresponding to each of the deep sub-parameters to be interpreted of the training samples based on the current cognitive layer characteristics of the training samples and the historical cognitive layer characteristics of the training samples; determining the deep attention weight distribution based on the difference between the spatial attention weight distribution of the training samples corresponding to each of the deep sub-parameters to be interpreted of the training samples and the spatiotemporal attention weight distribution of the training samples corresponding to each of the deep sub-parameters to be interpreted of the training samples. Loss, based on the spatial attention weight distribution of each training sample and the difference between each training sample autonomous driving control sub-parameter and the corresponding autonomous driving behavior label, determine the shallow attention loss; then aggregate the deep attention loss and the shallow attention loss to obtain the autonomous driving behavior prediction loss; then, judge whether the cognitive decision model has converged based on the autonomous driving behavior prediction loss. When it is determined that the cognitive decision model has converged, it can be determined that the model training is completed, and a trained cognitive decision model is obtained; when it is determined that the cognitive decision model has not converged, a round of updating of the cognitive decision model can be performed based on the autonomous driving behavior prediction loss according to the gradient descent method, and return to execute the step of obtaining the in-vehicle and out-of-vehicle monitoring data and autonomous driving behavior labels of the training samples to perform the next round of training.
可选地,所述基于各所述训练样本深层待解释子参数各自对应的训练样本空间注意力权重分布与各自对应的训练样本时空注意力权重分布之间的差异确定深层注意力损失的步骤包括:Optionally, the step of determining the deep attention loss based on the difference between the spatial attention weight distribution of the training samples corresponding to each of the deep sub-parameters to be interpreted of the training samples and the spatiotemporal attention weight distribution of the training samples corresponding to each of the deep sub-parameters to be interpreted of the training samples comprises:
将各所述训练样本空间注意力权重分布和各所述训练样本时空注意力权重分布输入深层注意力损失函数,确定深层注意力损失,其中,所述深层注意力损失函数为:Input the spatial attention weight distribution of each training sample and the spatiotemporal attention weight distribution of each training sample into the deep attention loss function to determine the deep attention loss, wherein the deep attention loss function is:
其中,DKL表示散度函数,为t行为时刻第i个的训练样本深层待解释子参数对应的训练样本空间注意力权重分布,为t行为时刻第i个的训练样本深层待解释子参数对应的训练样本时空注意力权重分布,λd为预设第二超参数,i∈[1,l]。Where D KL represents the divergence function, is the spatial attention weight distribution of the training sample corresponding to the deep-layer to-be-explained sub-parameter of the i-th training sample at behavior time t, is the spatiotemporal attention weight distribution of the training samples corresponding to the deep-layer to-be-explained sub-parameters of the i-th training sample at behavior time t, λd is the preset second hyperparameter, i∈[1,l].
可选地,所述将所述深层注意力损失和所述浅层注意力损失进行聚合,得到自动驾驶行为预测损失的步骤包括:Optionally, the step of aggregating the deep attention loss and the shallow attention loss to obtain the autonomous driving behavior prediction loss includes:
步骤E10,根据各所述训练样本浅层待解释子参数和所述训练样本当前认知层特征,生成至少一个训练样本浅层驾驶行为分步解释,根据各所述训练样本深层待解释子参数和所述训练样本当前认知层特征,生成至少一个训练样本深层驾驶行为分步解释;Step E10, generating at least one shallow driving behavior step-by-step explanation of the training sample according to the shallow sub-parameters to be explained of each training sample and the current cognitive layer characteristics of the training sample, and generating at least one deep driving behavior step-by-step explanation of the training sample according to the deep sub-parameters to be explained of each training sample and the current cognitive layer characteristics of the training sample;
步骤E20,将各所述训练样本浅层驾驶行为分步解释、各所述训练样本浅层待解释子参数、各所述训练样本深层驾驶行为分步解释和各所述训练样本深层待解释子参数进行聚合,得到训练样本驾驶行为解释;Step E20, aggregating the shallow driving behavior step-by-step explanations of each training sample, the shallow sub-parameters to be explained of each training sample, the deep driving behavior step-by-step explanations of each training sample, and the deep sub-parameters to be explained of each training sample to obtain the driving behavior explanations of the training samples;
步骤E30,获取上一时间步的历史训练样本驾驶行为解释、历史训练样本空间注意力权重分布、历史训练样本时空注意力权重分布,基于所述历史训练样本驾驶行为解释、所述历史训练样本空间注意力权重分布、所述历史训练样本时空注意力权重分布,以及当前时间步的训练样本驾驶行为解释、训练样本空间注意力权重分布和训练样本时空注意力权重分布,确定最小化负对数似然损失;Step E30, obtaining the historical training sample driving behavior explanation, the historical training sample spatial attention weight distribution, and the historical training sample spatiotemporal attention weight distribution of the previous time step, and determining the minimum negative log-likelihood loss based on the historical training sample driving behavior explanation, the historical training sample spatial attention weight distribution, and the historical training sample spatiotemporal attention weight distribution, as well as the training sample driving behavior explanation, the training sample spatial attention weight distribution, and the training sample spatiotemporal attention weight distribution of the current time step;
步骤E40,将所述深层注意力损失、所述浅层注意力损失和所述最小化负对数似然损失进行聚合,得到自动驾驶行为预测损失。Step E40, aggregating the deep attention loss, the shallow attention loss and the minimized negative log-likelihood loss to obtain the autonomous driving behavior prediction loss.
在本实施例中,需要说明的是,为了统一优化认知决策模型,使得认知决策模型整体损失最小,能够获得最佳的解释性能,将浅层注意力损失和深层注意力损失结合进行训练,并且通过最小化负对数似然函数进行模型微调。In this embodiment, it should be noted that in order to uniformly optimize the cognitive decision-making model so that the overall loss of the cognitive decision-making model is minimized and the best interpretation performance can be obtained, the shallow attention loss and the deep attention loss are combined for training, and the model is fine-tuned by minimizing the negative log-likelihood function.
作为一种示例,所述步骤E10-E40包括:根据各所述训练样本浅层待解释子参数和所述训练样本当前认知层特征,生成各所述训练样本浅层待解释子参数各自对应的训练样本浅层驾驶行为分步解释,根据各所述训练样本深层待解释子参数和所述训练样本当前认知层特征,生成各所述训练样本深层待解释子参数各自对应的训练样本深层驾驶行为分步解释;进而,将各所述训练样本浅层驾驶行为分步解释与各自对应的训练样本浅层待解释子参数聚合成训练样本驾驶行为浅层分步解释,并将各所述训练样本深层驾驶行为分步解释与各自对应的训练样本深层待解释子参数聚合成训练样本驾驶行为深层分步解释,进而将各所述训练样本驾驶行为浅层分步解释和各所述训练样本驾驶行为深层分步解释进行聚合,得到训练样本驾驶行为解释;进而,获取上一时间步的历史训练样本驾驶行为解释、历史训练样本空间注意力权重分布、历史训练样本时空注意力权重分布,将上一时间步的历史训练样本驾驶行为解释、历史训练样本空间注意力权重分布、历史训练样本时空注意力权重分布和当前时间步的训练样本驾驶行为解释、训练样本空间注意力权重分布和训练样本时空注意力权重分布,输入最小化负对数似然函数,计算最小化负对数似然损失;进而,将所述深层注意力损失、所述浅层注意力损失和所述最小化负对数似然损失进行聚合,得到自动驾驶行为预测损失。As an example, the steps E10-E40 include: generating a step-by-step explanation of the shallow driving behavior of the training sample corresponding to each shallow sub-parameter to be explained of the training sample and the current cognitive layer characteristics of the training sample according to the shallow sub-parameter to be explained of each training sample and the current cognitive layer characteristics of the training sample; generating a step-by-step explanation of the deep driving behavior of the training sample corresponding to each deep sub-parameter to be explained of the training sample according to the deep sub-parameter to be explained of each training sample and the current cognitive layer characteristics of the training sample; further, aggregating the step-by-step explanation of the shallow driving behavior of each training sample and the corresponding shallow sub-parameter to be explained of the training sample into a shallow step-by-step explanation of the driving behavior of the training sample, and aggregating the step-by-step explanation of the deep driving behavior of each training sample and the corresponding deep sub-parameter to be explained of the training sample into a deep step-by-step explanation of the driving behavior of the training sample, and further aggregating the step-by-step explanation of the deep driving behavior of each training sample and the corresponding deep sub-parameter to be explained of the training sample into a deep step-by-step explanation of the driving behavior of the training sample, and further aggregating the step-by-step explanation of the deep driving behavior of each training sample and the corresponding deep sub-parameter to be explained of the training sample The shallow step-by-step explanation of the driving behavior and the deep step-by-step explanation of the driving behavior of each training sample are aggregated to obtain the explanation of the driving behavior of the training sample; then, the driving behavior explanation of the historical training samples, the spatial attention weight distribution of the historical training samples, and the spatiotemporal attention weight distribution of the historical training samples at the previous time step are obtained, and the driving behavior explanation of the historical training samples, the spatial attention weight distribution of the historical training samples, and the spatiotemporal attention weight distribution of the historical training samples at the previous time step and the driving behavior explanation of the training samples, the spatial attention weight distribution of the training samples, and the spatiotemporal attention weight distribution of the training samples at the current time step are input into the minimized negative log-likelihood function to calculate the minimized negative log-likelihood loss; then, the deep attention loss, the shallow attention loss, and the minimized negative log-likelihood loss are aggregated to obtain the automatic driving behavior prediction loss.
在一种可实施的方式中,所述自动驾驶行为预测损失的损失函数可以表示为:In one practicable manner, the loss function of the autonomous driving behavior prediction loss can be expressed as:
其中, in,
其中,为自动驾驶行为预测损失,表示浅层注意力损失,表示深层注意力损失,k表示当前时间步,k-1表示上一时间步,表示当前时间步的注意力权重分布,包括空间注意力权重分布和时空注意力权重分布,Jk表示当前时间步的驾驶行为解释,ok表示当前时间步的可解释性输出,包括和Jk,ok-1表示上一时间步的可解释性输出,表示给定上一时间步的可解释性输出的情况下确定当前时间步的可解释性输出的概率。in, Predicting losses for autonomous driving behaviors, represents shallow attention loss, represents the deep attention loss, k represents the current time step, k-1 represents the previous time step, represents the attention weight distribution of the current time step, including the spatial attention weight distribution and the spatiotemporal attention weight distribution, J k represents the driving behavior explanation of the current time step, and o k represents the interpretable output of the current time step, including and J k , ok-1 represents the interpretable output of the previous time step, Represents the probability of determining the interpretable output at the current time step given the interpretable output at the previous time step.
在本实施例中,通过深层解释机制,可以对较为复杂场景中用户难以理解的驾驶行为进行时间维度和空间维度的深层解释,为用户解释驾驶行为产生的前因后果,消除用户对驾驶行为的疑惑,提高了端到端感知决策架构的可解释性,减少了用户因未知而产生的疑惑、恐慌、不安等负面情绪,提高了用户体验感。In this embodiment, through the deep interpretation mechanism, the driving behavior that is difficult for users to understand in more complex scenarios can be deeply interpreted in the time dimension and space dimension, explaining the causes and consequences of the driving behavior to users, eliminating users' doubts about the driving behavior, improving the interpretability of the end-to-end perception decision architecture, reducing users' negative emotions such as doubt, panic, and anxiety caused by the unknown, and improving user experience.
实施例四Embodiment 4
进一步地,本申请实施例还提供一种自动驾驶控制装置,参照图5,所述自动驾驶控制装置上部署有认知决策模型,所述认知决策模型包括认知层和决策层,所述自动驾驶控制装置包括:Furthermore, an embodiment of the present application further provides an automatic driving control device. Referring to FIG. 5 , a cognitive decision model is deployed on the automatic driving control device. The cognitive decision model includes a cognitive layer and a decision layer. The automatic driving control device includes:
认知模块10,用于获取车内外监测数据,通过所述认知层从所述车内外监测数据中提取当前认知层特征;A cognitive module 10, used to obtain in-vehicle and out-vehicle monitoring data, and extract current cognitive layer features from the in-vehicle and out-vehicle monitoring data through the cognitive layer;
决策模块20,用于将所述当前认知层特征输入所述决策层,确定自动驾驶控制参数;A decision module 20, configured to input the current cognitive layer features into the decision layer to determine the automatic driving control parameters;
解释模块30,用于根据所述当前认知层特征和所述自动驾驶控制参数,生成驾驶行为解释。The explanation module 30 is used to generate a driving behavior explanation based on the current cognitive layer characteristics and the automatic driving control parameters.
可选地,所述自动驾驶控制参数包括多个自动驾驶控制子参数;Optionally, the autonomous driving control parameter includes a plurality of autonomous driving control sub-parameters;
所述解释模块30,还用于:The interpretation module 30 is further used for:
根据所述当前认知层特征,确定各所述自动驾驶控制子参数各自对应的目标决策特征;Determining target decision features corresponding to each of the autonomous driving control sub-parameters according to the current cognitive layer features;
将各所述自动驾驶控制子参数与各自对应的目标决策特征分别进行聚合,得到驾驶行为分步解释;Aggregating each of the autonomous driving control sub-parameters with the corresponding target decision features to obtain a step-by-step explanation of the driving behavior;
将各所述驾驶行为分步解释进行聚合,得到驾驶行为解释。The step-by-step explanations of the driving behavior are aggregated to obtain a driving behavior explanation.
可选地,所述当前认知层特征包括车外环境特征;所述目标决策特征包括浅层目标决策特征;Optionally, the current cognitive layer features include features of the vehicle's external environment; the target decision features include shallow target decision features;
所述解释模块30,还用于:The interpretation module 30 is further used for:
根据所述当前认知层特征,从各所述自动驾驶控制子参数中确定至少一个浅层待解释子参数,以及各所述浅层待解释子参数各自对应的空间注意力权重分布;Determine, according to the current cognitive layer characteristics, at least one shallow-layer sub-parameter to be explained from each of the autonomous driving control sub-parameters, and a spatial attention weight distribution corresponding to each of the shallow-layer sub-parameters to be explained;
根据所述空间注意力权重分布和所述车外环境特征确定各所述浅层待解释子参数各自对应的浅层目标决策特征。The shallow target decision features corresponding to each of the shallow sub-parameters to be interpreted are determined according to the spatial attention weight distribution and the external vehicle environment characteristics.
可选地,所述自动驾驶控制装置还包括模型训练模块,所述模型训练模块用于:Optionally, the automatic driving control device further includes a model training module, and the model training module is used to:
获取训练样本车内外监测数据和自动驾驶行为标签,通过所述认知层从所述训练样本车内外监测数据中提取训练样本当前认知层特征;Acquire the in-vehicle and out-vehicle monitoring data and the autonomous driving behavior labels of the training samples, and extract the current cognitive layer features of the training samples from the in-vehicle and out-vehicle monitoring data of the training samples through the cognitive layer;
将所述训练样本当前认知层特征输入所述决策层,确定多个训练样本自动驾驶控制子参数;Inputting the current cognitive layer features of the training samples into the decision layer to determine a plurality of automatic driving control sub-parameters of the training samples;
根据所述训练样本当前认知层特征,从各所述训练样本自动驾驶控制子参数中确定至少一个训练样本浅层待解释子参数,以及各所述训练样本浅层待解释子参数各自对应的训练样本空间注意力权重分布;Determining at least one shallow sub-parameter to be explained of the training sample from each of the autonomous driving control sub-parameters of the training sample according to the current cognitive layer characteristics of the training sample, and the spatial attention weight distribution of the training samples corresponding to each of the shallow sub-parameters to be explained of the training sample;
基于各所述训练样本自动驾驶控制子参数、各所述自动驾驶行为标签以及各所述训练样本空间注意力权重分布确定浅层注意力损失,基于所述浅层注意力损失对所述认知决策模型进行迭代优化。The shallow attention loss is determined based on the autonomous driving control sub-parameters of each training sample, the autonomous driving behavior labels, and the spatial attention weight distribution of each training sample, and the cognitive decision-making model is iteratively optimized based on the shallow attention loss.
可选地,所述模型训练模块,还用于:Optionally, the model training module is further used to:
将各所述训练样本自动驾驶控制子参数、各所述自动驾驶行为标签以及各所述训练样本空间注意力权重分布输入浅层注意力损失函数,确定浅层注意力损失,其中,所述浅层注意力损失函数为:Input each of the training sample autonomous driving control sub-parameters, each of the autonomous driving behavior labels, and each of the training sample spatial attention weight distributions into a shallow attention loss function to determine a shallow attention loss, wherein the shallow attention loss function is:
其中,为浅层注意力损失,τt为t行为时刻的自动驾驶行为标签,为t行为时刻的训练样本自动驾驶控制子参数,H为熵函数,λs为预设第一超参数,为t行为时刻的训练样本空间注意力权重分布,t∈[t0,T]。in, is the shallow attention loss, τt is the autonomous driving behavior label at behavior time t, is the training sample automatic driving control sub-parameter at behavior time t, H is the entropy function, λ s is the preset first hyperparameter, is the spatial attention weight distribution of the training sample at behavior time t, t∈[t 0 ,T].
可选地,所述驾驶行为解释包括深层驾驶行为解释;所述目标决策特征包括深层目标决策特征;Optionally, the driving behavior explanation includes a deep driving behavior explanation; the target decision feature includes a deep target decision feature;
所述解释模块30,还用于:The interpretation module 30 is further used for:
根据所述当前认知层特征,从各所述自动驾驶控制子参数中确定至少一个深层待解释子参数;Determining at least one deep-level sub-parameter to be interpreted from each of the autonomous driving control sub-parameters according to the current cognitive layer characteristics;
获取当前时刻之前提取的历史认知层特征;Obtain the historical cognitive layer features extracted before the current moment;
采用时空注意力机制,基于所述当前认知层特征和所述历史认知层特征,提取各所述深层待解释子参数各自对应的深层目标决策特征。The spatiotemporal attention mechanism is adopted to extract the deep target decision features corresponding to each of the deep sub-parameters to be explained based on the current cognitive layer features and the historical cognitive layer features.
可选地,所述模型训练模块,还用于:Optionally, the model training module is further used to:
根据所述训练样本当前认知层特征,从各所述训练样本自动驾驶控制子参数中确定至少一个训练样本深层待解释子参数、以及各所述训练样本深层待解释子参数各自对应的训练样本空间注意力权重分布;Determining at least one training sample deep-layer to-be-explained sub-parameter from each of the training sample autonomous driving control sub-parameters according to the current cognitive layer characteristics of the training sample, and the spatial attention weight distribution of the training samples corresponding to each of the training sample deep-layer to-be-explained sub-parameters;
获取当前时刻之前提取的训练样本历史认知层特征,采用时空注意力机制,基于所述训练样本当前认知层特征和所述训练样本历史认知层特征,确定各所述训练样本深层待解释子参数各自对应的训练样本时空注意力权重分布;Obtain the historical cognitive layer features of the training samples extracted before the current moment, and use the spatiotemporal attention mechanism to determine the spatiotemporal attention weight distribution of the training samples corresponding to the deep-layer to-be-explained sub-parameters of each training sample based on the current cognitive layer features of the training samples and the historical cognitive layer features of the training samples;
基于各所述训练样本深层待解释子参数各自对应的训练样本空间注意力权重分布与各自对应的训练样本时空注意力权重分布之间的差异确定深层注意力损失,基于各所述训练样本空间注意力权重分布以及各所述训练样本自动驾驶控制子参数与各自对应的自动驾驶行为标签之间的差异,确定浅层注意力损失;Determine the deep attention loss based on the difference between the spatial attention weight distribution of the training samples corresponding to the deep sub-parameters to be interpreted of the training samples and the corresponding spatiotemporal attention weight distribution of the training samples; determine the shallow attention loss based on the difference between the spatial attention weight distribution of the training samples and the autonomous driving control sub-parameters of the training samples and the corresponding autonomous driving behavior labels;
将所述深层注意力损失和所述浅层注意力损失进行聚合,得到自动驾驶行为预测损失,基于所述自动驾驶行为预测损失对所述认知决策模型进行迭代优化。The deep attention loss and the shallow attention loss are aggregated to obtain the autonomous driving behavior prediction loss, and the cognitive decision model is iteratively optimized based on the autonomous driving behavior prediction loss.
可选地,所述模型训练模块,还用于:Optionally, the model training module is further used to:
将各所述训练样本空间注意力权重分布和各所述训练样本时空注意力权重分布输入深层注意力损失函数,确定深层注意力损失,其中,所述深层注意力损失函数为:Input the spatial attention weight distribution of each training sample and the spatiotemporal attention weight distribution of each training sample into the deep attention loss function to determine the deep attention loss, wherein the deep attention loss function is:
其中,DKL表示散度函数,为t行为时刻第i个的训练样本深层待解释子参数对应的训练样本空间注意力权重分布,为t行为时刻第i个的训练样本深层待解释子参数对应的训练样本时空注意力权重分布,λd为预设第二超参数,i∈[1,l]。Where D KL represents the divergence function, is the spatial attention weight distribution of the training sample corresponding to the deep-layer to-be-explained sub-parameter of the i-th training sample at behavior time t, is the spatiotemporal attention weight distribution of the training samples corresponding to the deep-layer to-be-explained sub-parameters of the i-th training sample at behavior time t, λd is the preset second hyperparameter, i∈[1,l].
可选地,所述模型训练模块,还用于:Optionally, the model training module is further used to:
根据各所述训练样本浅层待解释子参数和所述训练样本当前认知层特征,生成至少一个训练样本浅层驾驶行为分步解释,根据各所述训练样本深层待解释子参数和所述训练样本当前认知层特征,生成至少一个训练样本深层驾驶行为分步解释;Generate at least one shallow driving behavior step-by-step explanation of the training sample according to the shallow sub-parameters to be explained of each training sample and the current cognitive layer characteristics of the training sample, and generate at least one deep driving behavior step-by-step explanation of the training sample according to the deep sub-parameters to be explained of each training sample and the current cognitive layer characteristics of the training sample;
将各所述训练样本浅层驾驶行为分步解释、各所述训练样本浅层待解释子参数、各所述训练样本深层驾驶行为分步解释和各所述训练样本深层待解释子参数进行聚合,得到训练样本驾驶行为解释;Aggregating the shallow driving behavior step-by-step explanations of each training sample, the shallow sub-parameters to be explained of each training sample, the deep driving behavior step-by-step explanations of each training sample, and the deep sub-parameters to be explained of each training sample to obtain the driving behavior explanations of the training samples;
获取上一时间步的历史训练样本驾驶行为解释、历史训练样本空间注意力权重分布、历史训练样本时空注意力权重分布,基于所述历史训练样本驾驶行为解释、所述历史训练样本空间注意力权重分布、所述历史训练样本时空注意力权重分布,以及当前时间步的训练样本驾驶行为解释、训练样本空间注意力权重分布和训练样本时空注意力权重分布,确定最小化负对数似然损失;Obtain the historical training sample driving behavior explanation, the historical training sample spatial attention weight distribution, and the historical training sample spatiotemporal attention weight distribution of the previous time step, and determine the minimum negative log-likelihood loss based on the historical training sample driving behavior explanation, the historical training sample spatial attention weight distribution, the historical training sample spatiotemporal attention weight distribution, and the training sample driving behavior explanation, the training sample spatial attention weight distribution, and the training sample spatiotemporal attention weight distribution of the current time step;
将所述深层注意力损失、所述浅层注意力损失和所述最小化负对数似然损失进行聚合,得到自动驾驶行为预测损失。The deep attention loss, the shallow attention loss and the minimized negative log-likelihood loss are aggregated to obtain the autonomous driving behavior prediction loss.
可选地,所述解释模块30,还用于:Optionally, the interpretation module 30 is further configured to:
获取各所述自动驾驶控制子参数各自对应的分步解释权重;Obtaining a step-by-step explanation weight corresponding to each of the autonomous driving control sub-parameters;
基于各所述分步解释权重对各所述驾驶行为分步解释进行加权聚合,得到驾驶行为解释。The driving behavior explanation is obtained by weighting and aggregating the step-by-step explanations of the driving behavior based on the weights of the step-by-step explanations.
可选地,所述解释模块30,还用于:Optionally, the interpretation module 30 is further configured to:
获取上一时间步的注意力分布特征;Get the attention distribution features of the previous time step;
采用注意力机制,基于各所述驾驶行为分步解释以及所述上一时间步的注意力分布特征,确定各所述自动驾驶控制子参数各自对应的分步解释权重。An attention mechanism is adopted to determine the step-by-step interpretation weight corresponding to each of the autonomous driving control sub-parameters based on the step-by-step interpretation of each driving behavior and the attention distribution characteristics of the previous time step.
可选地,所述认知层包括用户特征提取模型;所述当前认知层特征包括用户需求特征,其中,所述用户需求特征包括用户隐式需求特征和用户显式需求特征;所述车内外监测数据包括用户操作数据和用户状态数据;Optionally, the cognitive layer includes a user feature extraction model; the current cognitive layer features include user demand features, wherein the user demand features include user implicit demand features and user explicit demand features; the in-vehicle and out-of-vehicle monitoring data include user operation data and user status data;
所述认知模块10,还用于:The cognitive module 10 is also used for:
通过所述用户特征提取模型从所述用户操作数据中提取用户显示需求特征,和/或,通过所述用户特征提取模型从所述用户状态数据中提取用户隐式需求特征。The user display demand features are extracted from the user operation data by using the user feature extraction model, and/or the user implicit demand features are extracted from the user status data by using the user feature extraction model.
可选地,所述认知层包括场景认知模型、地图构建模型、风险目标检测跟踪模型和运动预测模型;所述当前认知层特征包括环境认知特征,其中,所述环境认知特征包括场景认知特征、地图特征、风险目标检测跟踪特征和风险目标运动特征;Optionally, the cognitive layer includes a scene cognitive model, a map construction model, a risk target detection and tracking model, and a motion prediction model; the current cognitive layer features include environmental cognitive features, wherein the environmental cognitive features include scene cognitive features, map features, risk target detection and tracking features, and risk target motion features;
所述认知模块10,还用于:The cognitive module 10 is also used for:
通过所述场景认知模型从所述车内外监测数据中提取出场景认知特征,通过所述地图构建模型从所述车内外监测数据中提取出地图特征,并通过所述风险目标检测跟踪模型从所述车内外监测数据中提取出风险目标检测跟踪特征;Extracting scene recognition features from the vehicle interior and exterior monitoring data by using the scene recognition model, extracting map features from the vehicle interior and exterior monitoring data by using the map construction model, and extracting risk target detection and tracking features from the vehicle interior and exterior monitoring data by using the risk target detection and tracking model;
通过所述运动预测模型采用注意力机制,基于所述场景认知特征、所述地图特征和所述风险目标检测跟踪特征,进行风险目标运动预测,得到风险目标运动特征。The motion prediction model adopts the attention mechanism to perform risk target motion prediction based on the scene recognition features, the map features and the risk target detection and tracking features to obtain risk target motion features.
本发明提供的自动驾驶控制装置,采用上述实施例中的自动驾驶控制方法,解决了相关技术中端到端感知决策方案的可解释性较差的技术问题。与相关技术相比,本发明实施例提供的自动驾驶控制装置的有益与上述实施例提供的自动驾驶控制方法的有益相同,且该自动驾驶控制装置中的其他技术特征与上述实施例方法公开的特征相同,在此不做赘述。The automatic driving control device provided by the present invention adopts the automatic driving control method in the above embodiment, and solves the technical problem of poor interpretability of the end-to-end perception decision-making scheme in the related art. Compared with the related art, the benefits of the automatic driving control device provided by the embodiment of the present invention are the same as the benefits of the automatic driving control method provided by the above embodiment, and the other technical features in the automatic driving control device are the same as the features disclosed in the above embodiment method, which will not be repeated here.
实施例五Embodiment 5
进一步地,本申请实施例还提供一种自动驾驶控制装置,所述自动驾驶控制装置上部署有认知决策模型,所述认知决策模型包括认知层和决策层,所述认知层采用如上所述的自动驾驶控制方法进行预训练,所述自动驾驶控制装置包括:Furthermore, an embodiment of the present application also provides an automatic driving control device, on which a cognitive decision model is deployed, the cognitive decision model includes a cognitive layer and a decision layer, the cognitive layer is pre-trained using the automatic driving control method as described above, and the automatic driving control device includes:
第二获取模块,用于获取待预测车内外监测数据;The second acquisition module is used to acquire the vehicle internal and external monitoring data to be predicted;
第二认知模块,用于通过所述认知层从所述车内外监测数据中提取出待预测风险特征,其中,所述待预测风险特征包括待预测场景认知特征;A second cognitive module is used to extract risk features to be predicted from the vehicle internal and external monitoring data through the cognitive layer, wherein the risk features to be predicted include scene cognitive features to be predicted;
决策模块,用于将所述待预测风险特征输入所述决策层,确定自动驾驶控制参数。The decision module is used to input the risk characteristics to be predicted into the decision layer and determine the automatic driving control parameters.
可选地,所述认知层包括场景认知模型、地图构建模型、风险目标检测跟踪模型和运动预测模型;所述待预测风险特征还包括待预测地图特征、待预测风险目标检测跟踪特征和待预测风险目标运动特征;Optionally, the cognitive layer includes a scene cognitive model, a map construction model, a risk target detection and tracking model, and a motion prediction model; the risk features to be predicted also include map features to be predicted, risk target detection and tracking features to be predicted, and risk target motion features to be predicted;
所述第二认知模型,还用于:The second cognitive model is further used for:
通过所述场景认知模型从所述待预测车内外监测数据中提取出待预测场景认知特征,通过所述地图构建模型从所述待预测车内外监测数据中提取出待预测地图特征,通过所述风险目标检测跟踪模型从所述待预测车内外监测数据中提取出待预测风险目标检测跟踪特征;Extracting the scene recognition features to be predicted from the vehicle interior and exterior monitoring data to be predicted by the scene recognition model, extracting the map features to be predicted from the vehicle interior and exterior monitoring data to be predicted by the map construction model, and extracting the risk target detection and tracking features to be predicted from the vehicle interior and exterior monitoring data to be predicted by the risk target detection and tracking model;
通过所述运动预测模型采用注意力机制,基于所述待预测场景认知特征、所述待预测地图特征和所述待预测风险目标检测跟踪特征,进行风险目标运动预测,得到待预测风险目标运动特征。The motion prediction model adopts the attention mechanism to perform risk target motion prediction based on the cognitive features of the scene to be predicted, the map features to be predicted and the detection and tracking features of the risk target to be predicted, so as to obtain the motion features of the risk target to be predicted.
可选地,所述待预测风险目标检测跟踪特征包括至少一个风险目标检测跟踪子特征;所述运动预测模型包括注意力层、多层感知机层和预测层;Optionally, the risk target detection and tracking feature to be predicted includes at least one risk target detection and tracking sub-feature; the motion prediction model includes an attention layer, a multi-layer perceptron layer and a prediction layer;
所述第二认知模型,还用于:The second cognitive model is further used for:
将所述待预测场景认知特征、所述待预测地图特征和所述待预测风险目标检测跟踪特征输入所述注意力层,采用注意力机制,确定各所述风险目标检测跟踪子特征之间的风险目标交互特征、各所述风险目标检测跟踪子特征与所述待预测场景认知特征之间的场景交互特征以及各所述风险目标检测跟踪子特征与所述待预测地图特征之间的地图交互特征;Inputting the scene recognition feature to be predicted, the map feature to be predicted and the risk target detection and tracking feature to be predicted into the attention layer, and using the attention mechanism to determine the risk target interaction feature between each of the risk target detection and tracking sub-features, the scene interaction feature between each of the risk target detection and tracking sub-features and the scene recognition feature to be predicted, and the map interaction feature between each of the risk target detection and tracking sub-features and the map feature to be predicted;
将所述风险目标交互特征、所述场景交互特征以及所述地图交互特征输入多层感知机层,得到场景风险特征和运动查询特征;Inputting the risk target interaction feature, the scene interaction feature and the map interaction feature into a multi-layer perceptron layer to obtain a scene risk feature and a motion query feature;
将所述场景风险特征和所述运动查询特征输入所述预测层,进行风险目标运动预测,得到待预测风险目标运动特征。The scene risk feature and the motion query feature are input into the prediction layer to perform risk target motion prediction to obtain the risk target motion feature to be predicted.
本发明提供的自动驾驶控制装置,采用上述实施例中的自动驾驶控制方法,解决了相关技术中端到端感知决策方案的可解释性较差的技术问题。与相关技术相比,本发明实施例提供的自动驾驶控制装置的有益与上述实施例提供的自动驾驶控制方法的有益相同,且该自动驾驶控制装置中的其他技术特征与上述实施例方法公开的特征相同,在此不做赘述。The automatic driving control device provided by the present invention adopts the automatic driving control method in the above embodiment, and solves the technical problem of poor interpretability of the end-to-end perception decision-making scheme in the related art. Compared with the related art, the benefits of the automatic driving control device provided by the embodiment of the present invention are the same as the benefits of the automatic driving control method provided by the above embodiment, and the other technical features in the automatic driving control device are the same as the features disclosed in the above embodiment method, which will not be repeated here.
实施例六Embodiment 6
进一步地,本发明实施例提供一种电子设备,电子设备包括:至少一个处理器;以及,与至少一个处理器通信连接的存储器;其中,存储器存储有可被至少一个处理器执行的指令,指令被至少一个处理器执行,以使至少一个处理器能够执行上述实施例中的自动驾驶控制方法。Furthermore, an embodiment of the present invention provides an electronic device, which includes: at least one processor; and a memory communicatively connected to the at least one processor; wherein the memory stores instructions executable by the at least one processor, and the instructions are executed by the at least one processor so that the at least one processor can execute the autonomous driving control method in the above embodiment.
下面参考图6,其示出了适于用来实现本公开实施例的电子设备的结构示意图。本公开实施例中的电子设备可以包括但不限于诸如蓝牙耳机、移动电话、笔记本电脑、数字广播接收器、PDA(个人数字助理)、PAD(平板电脑)、PMP(便携式多媒体播放器)、车载终端(例如车载导航终端)等等的移动终端以及诸如数字TV、台式计算机等等的固定终端。图6示出的电子设备仅仅是一个示例,不应对本公开实施例的功能和使用范围带来任何限制。Referring to FIG6 below, a schematic diagram of the structure of an electronic device suitable for implementing the embodiment of the present disclosure is shown. The electronic device in the embodiment of the present disclosure may include, but is not limited to, mobile terminals such as Bluetooth headsets, mobile phones, laptop computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablet computers), PMPs (portable multimedia players), vehicle-mounted terminals (such as vehicle-mounted navigation terminals), etc., and fixed terminals such as digital TVs, desktop computers, etc. The electronic device shown in FIG6 is only an example and should not bring any limitation to the functions and scope of use of the embodiment of the present disclosure.
如图6所示,电子设备可以包括处理装置(例如中央处理器、图形处理器等),其可以根据存储在只读存储器(ROM)中的程序或者从存储装置加载到随机访问存储器(RAM)中的程序而执行各种适当的动作和处理。在RAM中,还存储有电子设备操作所需的各种程序和数组。处理装置、ROM以及RAM通过总线彼此相连。输入/输出(I/O)接口也连接至总线。As shown in Figure 6, the electronic device may include a processing device (e.g., a central processing unit, a graphics processing unit, etc.), which can perform various appropriate actions and processes according to a program stored in a read-only memory (ROM) or a program loaded from a storage device into a random access memory (RAM). In the RAM, various programs and arrays required for the operation of the electronic device are also stored. The processing device, ROM, and RAM are connected to each other via a bus. An input/output (I/O) interface is also connected to the bus.
通常,以下系统可以连接至I/O接口:包括例如触摸屏、触摸板、键盘、鼠标、图像传感器、麦克风、加速度计、陀螺仪等的输入装置;包括例如液晶显示器(LCD)、扬声器、振动器等的输出装置;包括例如磁带、硬盘等的存储装置;以及通信装置。通信装置可以允许电子设备与其他设备进行无线或有线通信以交换数组。虽然图中示出了具有各种系统的电子设备,但是应理解的是,并不要求实施或具备所有示出的系统。可以替代地实施或具备更多或更少的系统。Typically, the following systems can be connected to the I/O interface: input devices including, for example, a touch screen, a touchpad, a keyboard, a mouse, an image sensor, a microphone, an accelerometer, a gyroscope, etc.; output devices including, for example, a liquid crystal display (LCD), a speaker, a vibrator, etc.; storage devices including, for example, a magnetic tape, a hard disk, etc.; and communication devices. The communication device can allow the electronic device to communicate with other devices wirelessly or by wire to exchange arrays. Although the figure shows an electronic device with various systems, it should be understood that it is not required to implement or have all the systems shown. More or fewer systems may be implemented or have instead.
特别地,根据本公开的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本公开的实施例包括一种计算机程序产品,其包括承载在计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信装置从网络上被下载和安装,或者从存储装置被安装,或者从ROM被安装。在该计算机程序被处理装置执行时,执行本公开实施例的方法中限定的上述功能。In particular, according to an embodiment of the present disclosure, the process described above with reference to the flowchart can be implemented as a computer software program. For example, an embodiment of the present disclosure includes a computer program product, which includes a computer program carried on a computer-readable medium, and the computer program contains a program code for executing the method shown in the flowchart. In such an embodiment, the computer program can be downloaded and installed from a network through a communication device, or installed from a storage device, or installed from a ROM. When the computer program is executed by a processing device, the above-mentioned functions defined in the method of the embodiment of the present disclosure are executed.
本发明提供的电子设备,采用上述实施例中的自动驾驶控制方法,解决了相关技术中端到端感知决策方案的可解释性较差的技术问题。与相关技术相比,本发明实施例提供的电子设备的有益与上述实施例提供的自动驾驶控制方法的有益相同,且该电子设备中的其他技术特征与上述实施例方法公开的特征相同,在此不做赘述。The electronic device provided by the present invention adopts the automatic driving control method in the above embodiment, and solves the technical problem of poor interpretability of the end-to-end perception decision-making scheme in the related art. Compared with the related art, the benefits of the electronic device provided by the embodiment of the present invention are the same as the benefits of the automatic driving control method provided by the above embodiment, and the other technical features in the electronic device are the same as the features disclosed in the above embodiment method, which will not be repeated here.
应当理解,本公开的各部分可以用硬件、软件、固件或它们的组合来实现。在上述实施方式的描述中,具体特征、结构、材料或者特点可以在任何的一个或多个实施例或示例中以合适的方式结合。It should be understood that the various parts of the present disclosure can be implemented with hardware, software, firmware or a combination thereof. In the description of the above embodiments, specific features, structures, materials or characteristics can be combined in any one or more embodiments or examples in a suitable manner.
以上所述,仅为本发明的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本发明的保护范围之内。因此,本发明的保护范围应以所述权利要求的保护范围为准。The above is only a specific embodiment of the present invention, but the protection scope of the present invention is not limited thereto. Any person skilled in the art who is familiar with the technical field can easily think of changes or substitutions within the technical scope disclosed by the present invention, which should be included in the protection scope of the present invention. Therefore, the protection scope of the present invention should be based on the protection scope of the claims.
实施例七Embodiment 7
进一步地,本实施例提供一种计算机可读存储介质,具有存储在其上的计算机可读程序指令,计算机可读程序指令用于执行上述实施例中的自动驾驶控制方法。Furthermore, this embodiment provides a computer-readable storage medium having computer-readable program instructions stored thereon, and the computer-readable program instructions are used to execute the automatic driving control method in the above embodiment.
本发明实施例提供的计算机可读存储介质例如可以是U盘,但不限于电、磁、光、电磁、红外线、或半导体的系统、系统或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑磁盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本实施例中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、系统或者器件使用或者与其结合使用。计算机可读存储介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:电线、光缆、RF(射频)等等,或者上述的任意合适的组合。The computer-readable storage medium provided in the embodiment of the present invention may be, for example, a USB flash drive, but is not limited to electrical, magnetic, optical, electromagnetic, infrared, or semiconductor systems, systems or devices, or any combination thereof. More specific examples of computer-readable storage media may include, but are not limited to: an electrical connection with one or more wires, a portable computer disk, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disk read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination thereof. In the present embodiment, the computer-readable storage medium may be any tangible medium containing or storing a program, which may be used by or in combination with an instruction execution system, system or device. The program code contained on the computer-readable storage medium may be transmitted using any suitable medium, including but not limited to: wires, optical cables, RF (radio frequency), etc., or any suitable combination thereof.
上述计算机可读存储介质可以是电子设备中所包含的;也可以是单独存在,而未装配入电子设备中。The computer-readable storage medium may be included in the electronic device, or may exist independently without being installed in the electronic device.
上述计算机可读存储介质承载有一个或者多个程序,当上述一个或者多个程序被电子设备执行时,使得电子设备:获取车内外监测数据和场景标签,并获取所述场景标签对应的目标人类驾驶经验场景认知特征;通过所述认知层从所述车内外监测数据中提取出风险特征;根据所述风险特征和所述目标人类驾驶经验场景认知特征,对所述认知层进行迭代优化。The above-mentioned computer-readable storage medium carries one or more programs. When the above-mentioned one or more programs are executed by an electronic device, the electronic device: obtains in-vehicle and out-vehicle monitoring data and scene labels, and obtains target human driving experience scene cognition characteristics corresponding to the scene labels; extracts risk characteristics from the in-vehicle and out-vehicle monitoring data through the cognitive layer; and iteratively optimizes the cognitive layer according to the risk characteristics and the target human driving experience scene cognition characteristics.
或者,上述计算机可读存储介质承载有一个或者多个程序,当上述一个或者多个程序被电子设备执行时,使得电子设备:获取待预测车内外监测数据;通过所述认知层从所述车内外监测数据中提取出待预测风险特征,其中,所述待预测风险特征包括待预测场景认知特征;将所述待预测风险特征输入所述决策层,确定自动驾驶控制参数。Alternatively, the computer-readable storage medium carries one or more programs. When the one or more programs are executed by an electronic device, the electronic device: obtains the vehicle's internal and external monitoring data to be predicted; extracts risk features to be predicted from the vehicle's internal and external monitoring data through the cognitive layer, wherein the risk features to be predicted include scene cognitive features to be predicted; and inputs the risk features to be predicted into the decision layer to determine the automatic driving control parameters.
可以以一种或多种程序设计语言或其组合来编写用于执行本公开的操作的计算机程序代码,上述程序设计语言包括面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(LAN)或广域网(WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。Computer program code for performing the operations of the present disclosure may be written in one or more programming languages, or a combination thereof, including object-oriented programming languages, such as Java, Smalltalk, C++, and conventional procedural programming languages, such as "C" or similar programming languages. The program code may be executed entirely on the user's computer, partially on the user's computer, as a separate software package, partially on the user's computer and partially on a remote computer, or entirely on a remote computer or server. In cases involving a remote computer, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (e.g., through the Internet using an Internet service provider).
附图中的流程图和框图,图示了按照本发明各种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。The flow chart and block diagram in the accompanying drawings illustrate the possible architecture, function and operation of the system, method and computer program product according to various embodiments of the present invention. In this regard, each square box in the flow chart or block diagram can represent a module, a program segment or a part of a code, and the module, the program segment or a part of the code contains one or more executable instructions for realizing the specified logical function. It should also be noted that in some alternative implementations, the functions marked in the square box can also occur in a sequence different from that marked in the accompanying drawings. For example, two square boxes represented in succession can actually be executed substantially in parallel, and they can sometimes be executed in the opposite order, depending on the functions involved. It should also be noted that each square box in the block diagram and/or flow chart, and the combination of the square boxes in the block diagram and/or flow chart can be implemented with a dedicated hardware-based system that performs the specified function or operation, or can be implemented with a combination of dedicated hardware and computer instructions.
描述于本公开实施例中所涉及到的模块可以通过软件的方式实现,也可以通过硬件的方式来实现。其中,模块的名称在某种情况下并不构成对该单元本身的限定。The modules involved in the embodiments described in the present disclosure may be implemented by software or hardware, wherein the name of the module does not constitute a limitation on the unit itself in some cases.
本发明提供的计算机可读存储介质,存储有用于执行上述自动驾驶控制方法的计算机可读程序指令,解决了相关技术中端到端感知决策方案的可解释性较差的技术问题。与相关技术相比,本发明实施例提供的计算机可读存储介质的有益与上述实施例提供的自动驾驶控制方法的有益相同,在此不做赘述。The computer-readable storage medium provided by the present invention stores computer-readable program instructions for executing the above-mentioned automatic driving control method, which solves the technical problem of poor interpretability of the end-to-end perception decision-making scheme in the related art. Compared with the related art, the benefits of the computer-readable storage medium provided by the embodiment of the present invention are the same as the benefits of the automatic driving control method provided by the above-mentioned embodiment, which will not be repeated here.
以上仅为本申请的优选实施例,并非因此限制本申请的专利范围,凡是利用本申请说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本申请的专利处理范围内。The above are only preferred embodiments of the present application, and are not intended to limit the patent scope of the present application. Any equivalent structure or equivalent process transformation made using the contents of the present application specification and drawings, or directly or indirectly applied in other related technical fields, are also included in the patent processing scope of the present application.
Claims (16)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202311568377.2A CN117901892A (en) | 2023-11-22 | 2023-11-22 | Automatic driving control method, device, electronic device and storage medium |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202311568377.2A CN117901892A (en) | 2023-11-22 | 2023-11-22 | Automatic driving control method, device, electronic device and storage medium |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN117901892A true CN117901892A (en) | 2024-04-19 |
Family
ID=90684652
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202311568377.2A Pending CN117901892A (en) | 2023-11-22 | 2023-11-22 | Automatic driving control method, device, electronic device and storage medium |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN117901892A (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN119705504A (en) * | 2025-03-03 | 2025-03-28 | 吉林大学 | Causal knowledge-based interpretable automatic driving decision method |
-
2023
- 2023-11-22 CN CN202311568377.2A patent/CN117901892A/en active Pending
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN119705504A (en) * | 2025-03-03 | 2025-03-28 | 吉林大学 | Causal knowledge-based interpretable automatic driving decision method |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11449045B2 (en) | Artificial intelligence device and method of diagnosing malfunction using operation log and artificial intelligence model | |
| US20200050894A1 (en) | Artificial intelligence apparatus and method for providing location information of vehicle | |
| US20200042832A1 (en) | Artificial intelligence apparatus and method for updating artificial intelligence model | |
| WO2021073389A1 (en) | Vehicle automatic control method and lane change intention prediction network training method | |
| CN116991157B (en) | Automatic driving model with human expert driving capability, training method and vehicle | |
| US20230179662A1 (en) | Smart home device and method | |
| CN117585015A (en) | Automatic driving optimization methods, control methods, devices, electronic equipment and media | |
| US20210334461A1 (en) | Artificial intelligence apparatus and method for generating named entity table | |
| CN119636707B (en) | Control method and device for vehicle, electronic equipment and storage medium | |
| CN117901892A (en) | Automatic driving control method, device, electronic device and storage medium | |
| US11074814B2 (en) | Portable apparatus for providing notification | |
| Wang et al. | Human‐centric multimodal deep (HMD) traffic signal control | |
| CN117644877A (en) | Automatic driving optimization method, electronic device and storage medium | |
| US11392936B2 (en) | Exchange service robot and exchange service method using the same | |
| CN118551806A (en) | Automatic driving model, automatic driving method and device based on state node prediction | |
| US20200051571A1 (en) | Artificial intelligence device | |
| CN116558541B (en) | Model training method and device, and track prediction method and device | |
| CN117644863A (en) | Driving risk prediction method and device, electronic equipment and storage medium | |
| EP4481631A1 (en) | Edge-cloud collaboration system and method | |
| KR20190095190A (en) | Artificial intelligence device for providing voice recognition service and operating mewthod thereof | |
| WO2023051398A1 (en) | Security compensation method and apparatus, and storage medium and electronic device | |
| CN117657190A (en) | Automatic driving control method, device, electronic equipment and storage medium | |
| Alkorabi et al. | Deep learning algorithms for autonomous vehicle communications: Technical insights and open challenges | |
| US20210133561A1 (en) | Artificial intelligence device and method of operating the same | |
| US11971972B2 (en) | Artificial intelligence apparatus and method for controlling authority to use external device based on user identification using image recognition |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination |