[go: up one dir, main page]

CN111291590B - Driver fatigue detection method, driver fatigue detection device, computer equipment and storage medium - Google Patents

Driver fatigue detection method, driver fatigue detection device, computer equipment and storage medium Download PDF

Info

Publication number
CN111291590B
CN111291590B CN201811485916.5A CN201811485916A CN111291590B CN 111291590 B CN111291590 B CN 111291590B CN 201811485916 A CN201811485916 A CN 201811485916A CN 111291590 B CN111291590 B CN 111291590B
Authority
CN
China
Prior art keywords
value
image
frame
opening
opening degree
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811485916.5A
Other languages
Chinese (zh)
Other versions
CN111291590A (en
Inventor
彭斐
毛茜
何俏君
尹超凡
李彦琳
谷俊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Automobile Group Co Ltd
Original Assignee
Guangzhou Automobile Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Automobile Group Co Ltd filed Critical Guangzhou Automobile Group Co Ltd
Priority to CN201811485916.5A priority Critical patent/CN111291590B/en
Publication of CN111291590A publication Critical patent/CN111291590A/en
Application granted granted Critical
Publication of CN111291590B publication Critical patent/CN111291590B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/59Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
    • G06V20/597Recognising the driver's state or behaviour, e.g. attention or drowsiness
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Ophthalmology & Optometry (AREA)
  • Image Analysis (AREA)

Abstract

本发明涉及一种驾驶员疲劳检测方法、装置、计算机设备和存储介质,其方法包括:获取目标驾驶员的脸部视频,对所述脸部视频中的各帧脸部图像分别进行人眼开度检测,得到各帧所述脸部图像中的人眼开度值;根据各所述人眼开度值确定第一开度阈值和第二开度阈值,所述第一开度阈值大于所述第二开度阈值;根据各所述人眼开度值、所述第一开度阈值和所述第二开度阈值,统计所述人眼开度值小于或者等于所述第一开度阈值的第一图像帧数值,以及所述人眼开度值小于或者等于所述第二开度阈值的第二图像帧数值;若所述第一图像帧数值与所述第二开度阈值的比值大于预设的疲劳判定阈值,则判定所述目标驾驶员处于疲劳状态。采用本发明方案,可以提高疲劳检测结果的准确率。

Figure 201811485916

The invention relates to a driver fatigue detection method, device, computer equipment and storage medium. The method includes: acquiring a face video of a target driver, and performing human eye opening on each frame of face images in the face video respectively. degree detection to obtain the human eye opening value in each frame of the face image; determine a first opening degree threshold and a second opening degree threshold according to each of the human eye opening degree values, and the first opening degree threshold is greater than all the opening degree thresholds. the second opening degree threshold value; according to each of the human eye opening degree value, the first opening degree threshold value and the second opening degree threshold value, statistics that the human eye opening degree value is less than or equal to the first opening degree The first image frame value of the threshold, and the second image frame value of which the eye opening value is less than or equal to the second opening threshold; if the first image frame value and the second opening threshold are If the ratio is greater than the preset fatigue determination threshold, it is determined that the target driver is in a fatigue state. By adopting the scheme of the present invention, the accuracy of the fatigue detection result can be improved.

Figure 201811485916

Description

驾驶员疲劳检测方法、装置、计算机设备和存储介质Driver fatigue detection method, device, computer equipment and storage medium

技术领域technical field

本发明涉及图像处理技术领域,特别是涉及一种驾驶员疲劳检测方法、装置、计算机设备和存储介质。The present invention relates to the technical field of image processing, and in particular, to a driver fatigue detection method, device, computer equipment and storage medium.

背景技术Background technique

交通事故一直是人类面临的对生命财产安全威胁最严重的问题之一,其中大部分交通事故的发生都是由于驾驶员人为因素造成的。在车辆行驶过程中,驾驶员疲劳是造成恶性交通事故的重要原因之一,严重危害交通安全。Traffic accidents have always been one of the most serious threats to life and property safety faced by human beings, and most of the traffic accidents are caused by human factors of drivers. In the process of vehicle driving, driver fatigue is one of the important causes of serious traffic accidents, which seriously endangers traffic safety.

随着图像识别处理技术的发展,通过识别处理驾驶员驾驶过程中面部图像信息,判断驾驶员的疲劳状态并报警为预防交通事故的发生提供了一种新的解决方案。With the development of image recognition processing technology, it provides a new solution for preventing traffic accidents by identifying and processing the driver's facial image information during driving, judging the driver's fatigue state and giving an alarm.

传统的基于面部图像信息的驾驶员疲劳检测方式,是对人眼睁闭状态进行识别,最后通过连续帧的检测状态进行疲劳检测,这种方式,检测结果的准确率较低。The traditional driver fatigue detection method based on facial image information is to identify the open and closed state of human eyes, and finally perform fatigue detection through the detection state of continuous frames. In this way, the accuracy of the detection results is low.

发明内容SUMMARY OF THE INVENTION

基于此,有必要针对上述技术问题,提供一种能够提高检测准确率的驾驶员疲劳检测方法、装置、计算机设备和存储介质。Based on this, it is necessary to provide a driver fatigue detection method, device, computer equipment and storage medium that can improve the detection accuracy in response to the above technical problems.

一种驾驶员疲劳检测方法,所述方法包括:A driver fatigue detection method, the method comprises:

获取目标驾驶员的脸部视频,对所述脸部视频中的各帧脸部图像分别进行人眼开度检测,得到各帧所述脸部图像中的人眼开度值;Obtaining the facial video of the target driver, performing eye opening detection on each frame of the facial image in the facial video, respectively, to obtain the eye opening value in the facial image of each frame;

根据各所述人眼开度值确定第一开度阈值和第二开度阈值,所述第一开度阈值大于所述第二开度阈值;Determine a first opening threshold and a second opening threshold according to each of the human eye opening values, where the first opening threshold is greater than the second opening threshold;

根据各所述人眼开度值、所述第一开度阈值和所述第二开度阈值,统计所述人眼开度值小于或者等于所述第一开度阈值的第一图像帧数值,以及所述人眼开度值小于或者等于所述第二开度阈值的第二图像帧数值;According to each of the human eye opening degree value, the first opening degree threshold value and the second opening degree threshold value, count the first image frame values of which the human eye opening degree value is less than or equal to the first opening degree threshold value , and the second image frame value of which the eye opening value is less than or equal to the second opening threshold value;

若所述第一图像帧数值与所述第二图像帧数值的比值大于预设的疲劳判定阈值,则判定所述目标驾驶员处于疲劳状态If the ratio of the first image frame value to the second image frame value is greater than a preset fatigue determination threshold, it is determined that the target driver is in a fatigued state

一种驾驶员疲劳方法装置,所述装置包括:A driver fatigue method device, the device comprises:

检测模块,用于获取目标驾驶员的脸部视频,对所述脸部视频中的各帧脸部图像分别进行人眼开度检测,得到各帧所述脸部图像中的人眼开度值;The detection module is used to obtain the face video of the target driver, and perform eye opening detection on each frame of the face image in the face video, and obtain the eye opening value in each frame of the face image. ;

处理模块,用于根据各所述人眼开度值确定第一开度阈值和第二开度阈值,所述第一开度阈值大于所述第二开度阈值;a processing module, configured to determine a first opening threshold and a second opening threshold according to each of the human eye opening values, where the first opening threshold is greater than the second opening threshold;

统计模块,用于根据各所述人眼开度值、所述第一开度阈值和所述第二开度阈值,统计所述人眼开度值小于或者等于所述第一开度阈值的第一图像帧数值,以及所述人眼开度值小于或者等于所述第二开度阈值的第二图像帧数值;A statistics module, configured to count the number of eye openings less than or equal to the first opening threshold according to each of the human eye opening values, the first opening threshold and the second opening threshold. a first image frame value, and a second image frame value whose eye opening value is less than or equal to the second opening threshold;

判别模块,用于若所述第一图像帧数值与所述第二图像帧数值的比值大于预设的疲劳判定阈值,则判定所述目标驾驶员处于疲劳状态。A determination module, configured to determine that the target driver is in a fatigued state if the ratio of the first image frame value to the second image frame value is greater than a preset fatigue determination threshold.

一种计算机设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现以下步骤:A computer device, comprising a memory, a processor and a computer program stored in the memory and running on the processor, the processor implements the following steps when executing the computer program:

获取目标驾驶员的脸部视频,对所述脸部视频中的各帧脸部图像分别进行人眼开度检测,得到各帧所述脸部图像中的人眼开度值;Obtaining the facial video of the target driver, performing eye opening detection on each frame of the facial image in the facial video, respectively, to obtain the eye opening value in the facial image of each frame;

根据各所述人眼开度值确定第一开度阈值和第二开度阈值,所述第一开度阈值大于所述第二开度阈值;Determine a first opening threshold and a second opening threshold according to each of the human eye opening values, where the first opening threshold is greater than the second opening threshold;

根据各所述人眼开度值、所述第一开度阈值和所述第二开度阈值,统计所述人眼开度值小于或者等于所述第一开度阈值的第一图像帧数值,以及所述人眼开度值小于或者等于所述第二开度阈值的第二图像帧数值;According to each of the human eye opening degree value, the first opening degree threshold value and the second opening degree threshold value, count the first image frame values of which the human eye opening degree value is less than or equal to the first opening degree threshold value , and the second image frame value of which the eye opening value is less than or equal to the second opening threshold value;

若所述第一图像帧数值与所述第二图像帧数值的比值大于预设的疲劳判定阈值,则判定所述目标驾驶员处于疲劳状态。If the ratio of the first image frame value to the second image frame value is greater than a preset fatigue determination threshold, it is determined that the target driver is in a fatigued state.

一种计算机可读存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现以下步骤:A computer-readable storage medium on which a computer program is stored, and when the computer program is executed by a processor, the following steps are implemented:

获取目标驾驶员的脸部视频,对所述脸部视频中的各帧脸部图像分别进行人眼开度检测,得到各帧所述脸部图像中的人眼开度值;Obtaining the facial video of the target driver, performing eye opening detection on each frame of the facial image in the facial video, respectively, to obtain the eye opening value in the facial image of each frame;

根据各所述人眼开度值确定第一开度阈值和第二开度阈值,所述第一开度阈值大于所述第二开度阈值;Determine a first opening threshold and a second opening threshold according to each of the human eye opening values, where the first opening threshold is greater than the second opening threshold;

根据各所述人眼开度值、所述第一开度阈值和所述第二开度阈值,统计所述人眼开度值小于或者等于所述第一开度阈值的第一图像帧数值,以及所述人眼开度值小于或者等于所述第二开度阈值的第二图像帧数值;According to each of the human eye opening degree value, the first opening degree threshold value and the second opening degree threshold value, count the first image frame values of which the human eye opening degree value is less than or equal to the first opening degree threshold value , and the second image frame value of which the eye opening value is less than or equal to the second opening threshold value;

若所述第一图像帧数值与所述第二图像帧数值的比值大于预设的疲劳判定阈值,则判定所述目标驾驶员处于疲劳状态。If the ratio of the first image frame value to the second image frame value is greater than a preset fatigue determination threshold, it is determined that the target driver is in a fatigued state.

上述驾驶员疲劳检测方法、装置、计算机设备和存储介质,是根据脸部视频的各帧脸部图像中的各人眼开度值确定第一开度阈值和第二开度阈值,并基于第一开度阈值和第二开度阈值进行眼睛状态的区分,将基于人眼开度值大于第一开度阈值的第一图像帧数值和人眼开度值小于第二开度阈值的第二图像帧数值的比值,以及预设的疲劳判定阈值判定所述目标驾驶员处于疲劳状态,可以提升检测结果的准确度。The above-mentioned driver fatigue detection method, device, computer equipment and storage medium are to determine the first opening threshold and the second opening threshold according to the eye opening values in each frame of face image of the face video, and based on the first opening threshold and the second opening threshold. The first opening threshold and the second opening threshold are used to distinguish the state of the eyes, based on the first image frame value whose eye opening value is greater than the first opening threshold and the second image whose eye opening value is less than the second opening threshold. The ratio of the image frame values and the preset fatigue determination threshold determine that the target driver is in a fatigued state, which can improve the accuracy of the detection result.

附图说明Description of drawings

图1为一个实施例中驾驶员疲劳检测方法的应用环境图;1 is an application environment diagram of a driver fatigue detection method in one embodiment;

图2为一个实施例中驾驶员疲劳检测方法的流程示意图;2 is a schematic flowchart of a driver fatigue detection method in one embodiment;

图3为一个实施例中的人眼开度值的获取流程示意图;3 is a schematic flowchart of an acquisition process of a human eye opening value in one embodiment;

图4为一个实施例中的第一开度阈值和第二开度阈值的获取流程示意图;4 is a schematic flowchart of an acquisition process of a first opening degree threshold and a second opening degree threshold in one embodiment;

图5为一个实施例中的人脸特征图像的获取流程示意图;FIG. 5 is a schematic diagram of the acquisition process of the facial feature image in one embodiment;

图6为一个实施例中的眼睛特征点定位模型的训练流程示意图;6 is a schematic diagram of a training process of an eye feature point positioning model in one embodiment;

图7为一个实施例中的DPM特征提取原理图;7 is a schematic diagram of DPM feature extraction in one embodiment;

图8为一个实施例中的根滤波器(左)组件滤波器(中)高斯滤波后的2倍空间模型(右)的示意图;8 is a schematic diagram of a root filter (left) component filter (middle) Gaussian filtered 2x space model (right) in one embodiment;

图9为传统的Hog+SVM和一个实施例中的运用的DPM+Latent-SVM的效果对比(a)以及公式对比(b)示意图;9 is a schematic diagram of the effect comparison (a) and the formula comparison (b) of the DPM+Latent-SVM used in the traditional Hog+SVM and an embodiment;

图10为一个实施例中的联级迭代效果示意图;FIG. 10 is a schematic diagram of cascade iteration effect in one embodiment;

图11为一个实施例中的混合树模型对由于视点引起的拓扑变化进行编码的示意图;11 is a schematic diagram of a hybrid tree model encoding topology changes due to viewpoints in one embodiment;

图12为一个实施例中的人眼特征点定位结果示意图;FIG. 12 is a schematic diagram of the positioning result of human eye feature points in one embodiment;

图13为另一个实施例中驾驶员疲劳方法装置的结构框图;Fig. 13 is a structural block diagram of a driver fatigue method device in another embodiment;

图14为另一个实施例中计算机设备的内部结构图。FIG. 14 is an internal structure diagram of a computer apparatus in another embodiment.

具体实施方式Detailed ways

为使本发明的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本发明进行进一步的详细说明。应当理解,此处所描述的具体实施方式仅仅用以解释本发明,并不限定本发明的保护范围。In order to make the objectives, technical solutions and advantages of the present invention clearer, the present invention will be further described in detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are only used to explain the present invention, and do not limit the protection scope of the present invention.

需要说明的是,本发明的说明书、权利要求书以及说明书附图中的术语“第一”和“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后关系。应当理解这样使用的数据在适当情况下可以互换,以便这里描述的本发明的实施例能够除了在这里图示或描述的那些以外的顺序实施。It should be noted that the terms "first" and "second" in the description, claims and drawings of the present invention are used to distinguish similar objects, and are not necessarily used to describe a specific sequence or sequence relationship. It is to be understood that the data so used may be interchanged under appropriate circumstances so that the embodiments of the invention described herein can be practiced in sequences other than those illustrated or described herein.

本申请提供的驾驶员疲劳检测方法,可以应用于如图1所示的应用环境中。其中,红外摄像头采集驾驶员的视频信息,红外摄像头采集的视频信息可以输入终端中进行驾驶员疲劳方法。其中,红外摄像头较佳安装位置是汽车转向盘下方的转向柱上。红外摄像头可以通过有线或者无线的方式与终端进行通信。其中,终端102可以但不限于是各种个人计算机、笔记本电脑、智能手机、平板电脑、车载终端和便携式可穿戴设备。The driver fatigue detection method provided in this application can be applied to the application environment shown in FIG. 1 . Among them, the infrared camera collects the video information of the driver, and the video information collected by the infrared camera can be input into the terminal to carry out the driver fatigue method. Among them, the best installation position of the infrared camera is on the steering column below the steering wheel of the car. The infrared camera can communicate with the terminal in a wired or wireless manner. The terminal 102 may be, but is not limited to, various personal computers, notebook computers, smart phones, tablet computers, vehicle-mounted terminals and portable wearable devices.

在一个实施例中,如图2所示,提供了一种驾驶员疲劳检测方法,以该方法应用于终端为例进行说明,包括以下步骤:In one embodiment, as shown in FIG. 2 , a method for detecting driver fatigue is provided, which is illustrated by taking the method applied to a terminal as an example, including the following steps:

步骤S201:获取目标驾驶员的脸部视频,对所述脸部视频中的各帧脸部图像分别进行人眼开度检测,得到各帧所述脸部图像中的人眼开度值;Step S201: obtaining the face video of the target driver, and performing eye opening detection on each frame of the face image in the face video, to obtain the eye opening value in each frame of the face image;

这里,脸部视频是指对目标驾驶员面部进行拍摄得到的。Here, the face video is obtained by photographing the face of the target driver.

具体地,可以获取标驾驶员的一个检测周期内的脸部视频,对该脸部视频中的各帧脸部图像分别进行人眼开度检测,得到各帧所述脸部图像中的人眼开度值。其中,检测周期的大小可以根据实际需要设定。Specifically, the facial video of the target driver in one detection period can be obtained, and eye opening detection is performed on each frame of the facial image in the facial video to obtain the human eye in each frame of the facial image. opening value. The size of the detection period can be set according to actual needs.

步骤S202:根据各所述人眼开度值确定第一开度阈值和第二开度阈值,所述第一开度阈值大于所述第二开度阈值;Step S202: Determine a first opening threshold and a second opening threshold according to each of the human eye opening values, where the first opening threshold is greater than the second opening threshold;

一般地,第一开度阈值还小于目标驾驶员的人眼最大张开度值,第二开度阈值还大于0,第一开度阈值为用于判定眼睛是否处于完全睁开状态的阈值,第二开度阈值为用于判定眼睛是否处于闭合状态的阈值。Generally, the first opening threshold is smaller than the maximum opening degree of the target driver's eyes, the second opening threshold is greater than 0, the first opening threshold is a threshold for determining whether the eyes are in a fully opened state, and the second opening threshold is The opening threshold is a threshold for determining whether the eyes are in a closed state.

步骤S203:根据各所述人眼开度值、所述第一开度阈值和所述第二开度阈值,统计所述人眼开度值小于或者等于所述第一开度阈值的第一图像帧数值,以及所述人眼开度值小于或者等于所述第二开度阈值的第二图像帧数值;Step S203: According to each of the human eye opening value, the first opening degree threshold and the second opening degree threshold, count the first opening degree value of the human eye opening degree value less than or equal to the first opening degree threshold value. an image frame value, and a second image frame value whose eye opening value is less than or equal to the second opening threshold;

其中,人眼开度值大于所述第一开度阈值,表明眼睛处于完全睁开状态,人眼开度值小于所述第二开度阈值,表明眼睛处于闭合状态。第一图像帧数值是所述脸部视频中的人眼开度值小于或者等于所述第一开度阈值的图像帧数,第二图像帧数值是所述脸部视频中的人眼开度值小于或者等于所述第二开度阈值的图像帧数。Wherein, the human eye opening value is greater than the first opening degree threshold, indicating that the eyes are in a fully open state, and the human eye opening degree value is smaller than the second opening degree threshold, indicating that the eyes are in a closed state. The first image frame value is the number of image frames in which the human eye opening value in the face video is less than or equal to the first opening threshold value, and the second image frame value is the human eye opening degree in the face video. The number of image frames whose value is less than or equal to the second opening threshold.

具体地,可以根据各所述人眼开度值、所述第一开度阈值和所述第二开度阈值,统计一个检测周期内的第一图像帧数值和第二图像帧数值。Specifically, the first image frame value and the second image frame value in a detection period may be counted according to each of the human eye opening value, the first opening degree threshold and the second opening threshold.

步骤S204:若所述第一图像帧数值与所述第二开度阈值的比值大于预设的疲劳判定阈值,则判定所述目标驾驶员处于疲劳状态;Step S204: if the ratio of the first image frame value to the second opening threshold is greater than a preset fatigue determination threshold, determine that the target driver is in a fatigued state;

其中,疲劳判定阈值的大小可以根据实际情况设定。Among them, the size of the fatigue determination threshold can be set according to the actual situation.

上述驾驶员疲劳检测方法中,是获取目标驾驶员的脸部视频,对所述脸部视频中的各帧脸部图像分别进行人眼开度检测,得到各帧所述脸部图像中的人眼开度值,根据各所述人眼开度值确定第一开度阈值和第二开度阈值,所述第一开度阈值大于所述第二开度阈值,根据各所述人眼开度值、所述第一开度阈值和所述第二开度阈值,统计所述人眼开度值小于或者等于所述第一开度阈值的第一图像帧数值,以及所述人眼开度值小于或者等于所述第二开度阈值的第二图像帧数值,若所述第一图像帧数值与所述第二图像帧数值的比值大于预设的疲劳判定阈值,则判定所述目标驾驶员处于疲劳状态。本实施例方案中,基于第一开度阈值和第二开度阈值进行眼睛状态的区分,将基于人眼开度值大于第一开度阈值的第一图像帧数值和人眼开度值小于第二开度阈值的第二图像帧数值的比值,以及预设的疲劳判定阈值判定所述目标驾驶员处于疲劳状态,可以提升检测结果的准确度。In the above-mentioned driver fatigue detection method, the facial video of the target driver is obtained, and eye opening detection is performed on each frame of the facial image in the facial video, and the person in each frame of the facial image is obtained. Eye opening degree value, a first opening degree threshold and a second opening degree threshold are determined according to each of the human eye opening degree values, the first opening degree threshold value is greater than the second opening degree threshold value, and a first opening degree threshold value is determined according to each of the human eye opening degree thresholds. degree value, the first opening degree threshold value and the second opening degree threshold value, count the first image frame value whose eye opening degree value is less than or equal to the first opening degree threshold value, and the human eye opening degree value If the ratio of the first image frame value to the second image frame value is greater than the preset fatigue determination threshold, then determine the target The driver is fatigued. In the solution of this embodiment, the eye state is distinguished based on the first opening degree threshold and the second opening degree threshold, and the first image frame value based on the human eye opening degree value greater than the first opening degree threshold value and the human eye opening degree value less than The ratio of the second image frame value to the second opening threshold and the preset fatigue determination threshold to determine that the target driver is in a fatigued state can improve the accuracy of the detection result.

在其中一个实施例中,为了排除人眼与摄像头之间距离的变化对人眼开度计算造成的影响,提供一种对人眼开度值进行归一化的处理方式。如图3所示,具体地,上述的对所述脸部视频中的各帧脸部图像分别进行人眼开度检测,得到各帧所述脸部图像中的人眼开度值,可以包括:In one of the embodiments, in order to exclude the influence of the change of the distance between the human eye and the camera on the calculation of the human eye opening, a processing method for normalizing the human eye opening value is provided. As shown in FIG. 3 , specifically, the above-mentioned eye opening detection is performed on each frame of the face image in the face video, and the eye opening value in each frame of the face image is obtained, which may include: :

步骤S301:对各帧所述脸部图像分别进行眼部特征点定位,得到各帧所述脸部图像中的眼部特征点;Step S301: Perform eye feature point positioning on each frame of the face image to obtain the eye feature point in each frame of the face image;

步骤S302:根据各帧所述脸部图像中的眼部特征点,分别确定各帧所述脸部图像中的人眼瞳距值和人眼开度原始值;Step S302: According to the eye feature points in the facial images of each frame, respectively determine the interpupillary distance value of the human eye and the original value of the human eye opening in the facial image of each frame;

具体地,可以根据各帧所述脸部图像中的眼部特征点,分别确定各帧所述脸部图像中的正对瞳孔处的上眼睑的标定点和下眼睑的标定点,根据上眼睑的标定点和下眼睑的标定点来计算人眼开度原始值,其中,人眼开度原始值可以等于上眼睑的标定点和下眼睑的标定点之间的距离。Specifically, the calibration point of the upper eyelid and the calibration point of the lower eyelid at the position facing the pupil in each frame of the facial image may be determined respectively according to the eye feature points in the facial image of each frame, The calibration point and the calibration point of the lower eyelid are used to calculate the original value of the human eye opening, wherein the original value of the human eye opening may be equal to the distance between the calibration point of the upper eyelid and the calibration point of the lower eyelid.

步骤S303:根据各帧所述脸部图像中的人眼瞳距值和人眼开度原始值,分别确定各帧所述脸部图像中的人眼开度值。Step S303 : According to the interpupillary distance value of the human eye and the original value of the human eye opening in the facial image of each frame, determine the human eye opening value in each frame of the facial image respectively.

具体地,可以根据

Figure GDA0002899787390000061
确定各帧所述脸部图像中的人眼开度值,其中,Hi表示第i帧所述脸部图像中的人眼开度值,hi表示第i帧所述脸部图像中的人眼开度原始值,li表示第i帧所述脸部图像中的人眼瞳距值,C表示修正参数,该修正参数的大小可以根据情况设定或者调整。Specifically, according to
Figure GDA0002899787390000061
Determine the eye opening value in each frame of the face image, where H i represents the human eye opening value in the ith frame of the face image, and hi represents the ith frame of the face image. The original value of the human eye opening, li represents the interpupillary distance value of the human eye in the face image in the ith frame, and C represents a correction parameter, and the size of the correction parameter can be set or adjusted according to the situation.

采用本实施例中的方案,可以有效地排除人眼与摄像头之间距离的变化对人眼开度计算造成的影响。By adopting the solution in this embodiment, the influence of the change of the distance between the human eye and the camera on the calculation of the opening degree of the human eye can be effectively excluded.

在其中一个实施例中,如图4所示,上述的根据各所述人眼开度值确定第一开度阈值和第二开度阈值,包括:In one embodiment, as shown in FIG. 4 , the above-mentioned determination of the first opening threshold and the second opening threshold according to each of the human eye opening values includes:

步骤S401:根据各所述人眼开度值确定人眼最大张开度值;Step S401: Determine the maximum opening degree value of the human eye according to each of the human eye opening degree values;

具体地,可以比较各所述人眼开度值的大小,得到各所述人眼开度值中的最大值,将最大值作为人眼最大张开度值。也可以将各所述人眼开度值按照从大到小的顺序进行排序,对获取排序中的前N个人眼开度值,对这N个人眼开度值取均值,该均值作为人眼最大张开度值。Specifically, the magnitude of each of the human eye opening degree values may be compared to obtain the maximum value among the various human eye opening degree values, and the maximum value may be used as the maximum human eye opening degree value. It is also possible to sort each of the human eye opening values in descending order, obtain the top N eye opening values in the sorting, and take the average of these N human eye opening values, and the average value is used as the human eye. Maximum opening value.

步骤S402:将所述最大张开度值分别与预设的第一比例系数和第二比例系数相乘,得到所述第一开度阈值和所述第二开度阈值;Step S402: Multiply the maximum opening degree value by a preset first proportional coefficient and a second proportional coefficient, respectively, to obtain the first opening degree threshold and the second opening degree threshold;

其中,第一比例系数大于第二比例系数,且一般地,第一比例系数和第二比例系数均为大于0且小于1的值,第一比例系数和第二比例系数的大小可以根据实际需要选取,较佳地,第一比例系数为0.8,第二比例系数为0.2。The first proportional coefficient is greater than the second proportional coefficient, and generally, both the first proportional coefficient and the second proportional coefficient are values greater than 0 and less than 1, and the sizes of the first proportional coefficient and the second proportional coefficient can be based on actual needs Select, preferably, the first proportional coefficient is 0.8, and the second proportional coefficient is 0.2.

本实施例中,通过将所述最大张开度值分别与预设的第一比例系数和第二比例系数相乘得到所述第一开度阈值和所述第二开度阈值,算法易于实现,且第一开度阈值和第二开度阈值均基于人眼最大张开度值确定的,人眼最大张开度值又是由各人眼开度值确定的,可以进一步提升检测精度。In this embodiment, the first opening threshold and the second opening threshold are obtained by multiplying the maximum opening degree value by the preset first proportional coefficient and the second proportional coefficient respectively, and the algorithm is easy to implement, and The first opening degree threshold and the second opening degree threshold are both determined based on the maximum opening degree value of the human eye, and the maximum opening degree value of the human eye is determined by the opening degree value of each human eye, which can further improve the detection accuracy.

在其中一个实施例中,如图5所示,上述的对各帧所述脸部图像分别进行人脸特征点定位,获得各人脸特征图像,可以包括:In one embodiment, as shown in FIG. 5 , the above-mentioned positioning of the facial feature points of each frame of the facial image to obtain each facial feature image may include:

步骤S501:提取第一DPM特征图,所述第一DPM特征图为当前脸部图像的DPM特征图,所述当前脸部图像为任意一帧脸部图像;Step S501: extracting a first DPM feature map, where the first DPM feature map is a DPM feature map of a current facial image, and the current facial image is any frame of facial images;

步骤S502:对所述第一DPM特征图进行采样处理,提取第二DPM特征图,所述第二DPM特征图为对所述第一DPM特征图进行采样处理后的图像的DPM特征图;Step S502: performing sampling processing on the first DPM feature map, and extracting a second DPM feature map, where the second DPM feature map is a DPM feature map of an image obtained by sampling the first DPM feature map;

步骤S503:将所述第一DPM特征图,用预先训练的根滤波器进行卷积运算,得到所述根滤波器的响应图;Step S503: performing a convolution operation on the first DPM feature map with a pre-trained root filter to obtain a response map of the root filter;

步骤S504:将N倍的所述第二DPM特征图,用预先训练的组件滤波器进行卷积运算,得到所述组件滤波器的响应图,所述组件滤波器的分辨率为所述根滤波器的分辨率的N倍,N为正整数;Step S504: Perform a convolution operation on N times the second DPM feature map with a pre-trained component filter to obtain a response map of the component filter, and the resolution of the component filter is the root filter. N times the resolution of the device, N is a positive integer;

步骤S505:根据所述根滤波器的响应图和所述组件滤波器的响应图,得到目标响应图;Step S505: obtain a target response diagram according to the response diagram of the root filter and the response diagram of the component filter;

步骤S506:根据目标响应图获取当前人脸特征图像。Step S506: Acquire the current face feature image according to the target response map.

可以将对各帧所述脸部图像分别作为本实施例中的当前脸部图像,分别采用上述步骤S501~S506进行人脸特征点定位,获得各帧所述脸部图像对应的人脸特征图像。The facial image of each frame can be regarded as the current facial image in this embodiment, and the above steps S501 to S506 can be used to locate the facial feature points, and the facial feature image corresponding to each frame of the facial image can be obtained. .

本实施例中,通过采用采用DPM目标检测算法进行人脸检测,算法的检测准确率被提升,可以同时降低误检率和漏检率。In this embodiment, by adopting the DPM target detection algorithm for face detection, the detection accuracy of the algorithm is improved, and the false detection rate and the missed detection rate can be reduced at the same time.

在其中一个实施例中,对各帧所述脸部图像分别进行眼部特征点定位,得到各帧所述脸部图像中的眼部特征点,可以包括:对各帧所述脸部图像分别进行人脸特征点定位,获得各人脸特征图像;将各所述人脸特征图像分别输入预设的眼睛特征点定位模型,得到各帧所述脸部图像中的眼部特征点。In one embodiment, performing eye feature point positioning on each frame of the face image to obtain the eye feature point in each frame of the face image may include: Perform facial feature point location to obtain each facial feature image; input each of the facial feature images into a preset eye feature point location model respectively to obtain the eye feature points in each frame of the facial image.

在其中一个实施例中,如图6所示,上述的眼睛特征点定位模型的训练过程,可以包括:In one embodiment, as shown in FIG. 6 , the training process of the above-mentioned eye feature point positioning model may include:

步骤S601:获取目标图像各个像素点的像素值和各所述像素点出的特征向量;Step S601: Obtain the pixel value of each pixel point of the target image and the feature vector obtained by each of the pixel points;

本实施例中,使用的模型可以是基于混合树和共享的部件V池。将每个面部地标作为一个部分进行建模,并使用全局混合来捕获由于视点引起的拓扑变化。In this embodiment, the model used may be based on a hybrid tree and a shared component V pool. Model each facial landmark as a part and use global blending to capture topological changes due to viewpoints.

步骤S602:根据所述像素值和所述特征向量,配置树结构局部模型,并确定在L部分的得分函数;Step S602: according to the pixel value and the feature vector, configure the local model of the tree structure, and determine the score function in the L part;

其中,得分函数为S(I,L,m)=Appm(I,L)+Shapem(L)+αm

Figure GDA0002899787390000081
Figure GDA0002899787390000082
I表示目标图像,li=(xi,yi)表示所述目标图像的第i个像素点的像素值,w表示部分模型,m表示树结构是混合型,部分模型是指将所述目标图像中的每个面部特征点分别作为一个部分进行建模得到,a、b、c和d表示弹性参数,α表示混合偏置标量;Among them, the score function is S(I,L,m)=App m (I,L)+Shape m (L)+α m ,
Figure GDA0002899787390000081
Figure GDA0002899787390000082
I represents the target image, li =(x i , y i ) represents the pixel value of the ith pixel of the target image, w represents a partial model, m represents a hybrid tree structure, and a partial model refers to the Each facial feature point in the target image is modeled as a part, a, b, c, and d represent elastic parameters, and α represents a hybrid bias scalar;

本实施例中,

Figure GDA0002899787390000083
求出了在li位置上i处(第i个像素点处)的模板
Figure GDA0002899787390000084
之和,φ(I,li)表示了目标图像li像素处的特征向量。In this embodiment,
Figure GDA0002899787390000083
Find the template at position i (at the i-th pixel) at position li
Figure GDA0002899787390000084
The sum, φ (I, li ) represents the feature vector at the pixel of the target image li.

本实施例中,

Figure GDA0002899787390000085
表示的是混合类型特定空间L排列的排列得分,其中dx=xi-xj和dy=yi-yj表示第i部分到第j部分的位移。公式中的每个参数(指a,b,c,d)可以被解释为不同部分之间的空间约束。In this embodiment,
Figure GDA0002899787390000085
is the permutation score of the hybrid type specific spatial L permutation, where dx = x i -x j and dy = y i -y j represent the displacement of the i-th part to the j-th part. Each parameter (referring to a, b, c, d) in the formula can be interpreted as a spatial constraint between different parts.

步骤S603:通过求取使所述得分函数得到最大值的L和m的值的方式,得到各混合型的各个部分的最佳配置参数;Step S603: Obtain the optimal configuration parameters of each part of each hybrid type by obtaining the values of L and m that enable the score function to obtain the maximum value;

具体地,可以枚举所有混合型,对于每个混合型找到各部分的最佳配置参数。Specifically, all hybrid types can be enumerated, and for each hybrid type, the best configuration parameters of each part can be found.

步骤S604:建立训练样本集,所述训练样本集包括设定有标签的正样本和负样本,所述正样本为含有人脸的图像,所述负样本为不含有人脸的图像;Step S604: establishing a training sample set, the training sample set includes positive samples and negative samples set with labels, the positive samples are images containing human faces, and the negative samples are images that do not contain human faces;

具体地,假定一个全监督场景,在这个场景中有含有人脸的正样本和混合标签以及不含有人脸的负样本。可以用结构预测框架有区别地学习形状参数和外观参数。Specifically, a fully supervised scene is assumed, in which there are positive samples with faces and mixed labels and negative samples without faces. Shape parameters and appearance parameters can be learned differentially with a structure prediction framework.

步骤S605:根据所述部分模型、所述弹性参数以及所述混合偏置标量构建目标向量,根据所述目标向量修改所述得分函数;Step S605: Construct a target vector according to the partial model, the elastic parameter and the mixed bias scalar, and modify the score function according to the target vector;

具体地,将部分模型w、弹性参数(a,b,c,d)以及混合偏置标量α全部放进一个向量β中,将上述的得分函数修改成如下形式:S(I,z)=β·Φ(I,z)。向量Φ(I,z)是稀疏的,在与混合型m相对应的单个区间中具有非零项。Specifically, part of the model w, elastic parameters (a, b, c, d) and mixed bias scalar α are all put into a vector β, and the above score function is modified into the following form: S(I,z)= β·Φ(I,z). The vector Φ(I,z) is sparse, with non-zero entries in a single interval corresponding to the mixture m.

步骤S606:根据所述训练样本集、所述最佳配置参数、修改后的所述得分函数以及预先定义的目标预测函数,学习得到所述眼睛特征点定位模型;Step S606: Learning to obtain the eye feature point positioning model according to the training sample set, the optimal configuration parameter, the modified score function and the predefined target prediction function;

其中,学习得到的眼睛特征点定位模型为:Among them, the learned eye feature point localization model is:

Figure GDA0002899787390000091
Figure GDA0002899787390000091

Figure GDA0002899787390000092
Figure GDA0002899787390000092

Figure GDA0002899787390000093
Figure GDA0002899787390000093

Figure GDA0002899787390000094
Figure GDA0002899787390000094

其中,β表示所述目标向量,zn={Ln,mn},C代表目标函数的惩罚项系数,ξn表示第n个样本的惩罚项,pos和neg分别代表正样本和负样本,K代表所述目标向量的个数,k代表对应的所述目标向量的编号。Among them, β represents the target vector, z n ={L n ,m n }, C represents the penalty term coefficient of the objective function, ξ n represents the penalty term of the nth sample, pos and neg represent positive samples and negative samples, respectively , K represents the number of the target vector, and k represents the number of the corresponding target vector.

本实施例中,使用机器学习算法定位人脸特征点与并分别对眼睛特征点进行定位,不但定位精度非常高,且对光照与姿态有很强的泛化能力,可以提高眼睛开闭程度计算的精度。In this embodiment, the machine learning algorithm is used to locate the face feature points and the eye feature points respectively, which not only has a very high positioning accuracy, but also has a strong generalization ability for illumination and posture, which can improve the calculation of the degree of eye opening and closing. accuracy.

为了便于理解本发明的方案,以下以一个较佳实施例对本发明方案进行详细阐述。In order to facilitate the understanding of the solution of the present invention, the solution of the present invention is described in detail below with a preferred embodiment.

在该实施例中的驾驶员疲劳检测方法包括以下步骤:第一步:视频信息输入;第二步:人脸检测;第三步:人脸特征点定位;第四步:人眼特征点定位;第五步:眨眼检测—计算眼睛的开度,疲劳检测分析。The driver fatigue detection method in this embodiment includes the following steps: the first step: video information input; the second step: face detection; the third step: facial feature point location; the fourth step: human eye feature point location ; Step 5: Blink detection - calculation of eye opening, fatigue detection and analysis.

第一步:采集视频信息。用于视频信息输入的单目红外摄像头(安装于转向盘下方的转向柱上)实时输入驾驶员开车过程中脸部状态信息(图像)。视频输入的频率为30Hz,每帧图像大小为1280×1080像素。其中,红外摄像头能适应车内不同的光线情况,准确捕捉到驾驶员头部姿态信息和面部信息。Step 1: Collect video information. The monocular infrared camera for video information input (installed on the steering column under the steering wheel) inputs the driver's face status information (image) in real time while driving. The frequency of video input is 30Hz, and the size of each frame is 1280×1080 pixels. Among them, the infrared camera can adapt to different light conditions in the car, and accurately capture the driver's head posture information and facial information.

第二步:人脸检测。对输入的视频的每一帧图像,本实施例采用DPM(DeformablePart Model)目标检测算法进行人脸检测。DPM算法中应用到了HOG算法中的部分原理:首先是将图片灰度化;然后,如(1)式,采用Gamma校正法对输入图像进行颜色空间的标准化(归一化):The second step: face detection. For each frame image of the input video, this embodiment uses a DPM (Deformable Part Model) target detection algorithm to perform face detection. Part of the principle of the HOG algorithm is applied to the DPM algorithm: first, the image is grayed; then, as in formula (1), the color space of the input image is standardized (normalized) by the Gamma correction method:

I(x,y)=I(x,y)gamma (1)I(x,y)=I(x,y) gamma (1)

其中,gamma的取值看具体情况(例如可以取1/2),如此能够有效地降低图像局部的阴影和光照变化;接下来进行梯度计算,梯度反应的是相邻的像素之间的变化,相邻像素之间变化比较平坦,则梯度较小,反之梯度大,模拟图象f(x,y)任意一点像素(x,y)的梯度是一个矢量:Among them, the value of gamma depends on the specific situation (for example, it can be taken as 1/2), which can effectively reduce the local shadow and illumination changes of the image; next, the gradient calculation is performed, and the gradient reflects the change between adjacent pixels. The change between adjacent pixels is relatively flat, the gradient is small, otherwise the gradient is large, and the gradient of any pixel (x, y) of the simulated image f(x, y) is a vector:

Figure GDA0002899787390000095
Figure GDA0002899787390000095

Figure GDA0002899787390000096
Figure GDA0002899787390000096

其中,Gx是沿x方向上的梯度,Gy是沿y方向上的梯度,梯度的幅值及方向角可用如下公式表示:Among them, G x is the gradient along the x direction, G y is the gradient along the y direction, and the magnitude and direction angle of the gradient can be expressed by the following formula:

Figure GDA0002899787390000101
Figure GDA0002899787390000101

数字图像中的像素点使用差分来计算的:Pixels in a digital image are computed using difference:

Figure GDA0002899787390000102
Figure GDA0002899787390000102

因为使用简单的一维离散微分模板[-1,0,1]进行的梯度运算得到的检测效果是最好的,所以采用的计算公式如下:Because the gradient operation using a simple one-dimensional discrete differential template [-1,0,1] has the best detection effect, the calculation formula used is as follows:

Figure GDA0002899787390000103
值,其梯度的幅值及方向计算公式如下:
Figure GDA0002899787390000103
The formula for calculating the magnitude and direction of the gradient is as follows:

Figure GDA0002899787390000104
Figure GDA0002899787390000104

Figure GDA0002899787390000105
Figure GDA0002899787390000105

然后,对于整个目标图片,将其分成互不重叠、大小相同的细胞单元(cell),然后计算出每个细胞单元的梯度大小和方向。DPM保留了HOG图的细胞单元,然后将图片上某细胞单元(图7上是8x8的细胞单元)与其对角线邻域的四个细胞进行归一化操作。提取有符号的HOG梯度,0-360度将产生18个梯度向量,提取无符号的HOG梯度,0-180度将产生9个梯度向量。DPM只提取无符号特征,将产生4*9=36维特征,将行和列分别相加形成13个特征向量(如7图所示9列相加,4行相加),为了进一步提高精度,将提取的18维有符号的梯度特征也加进来(如图所示18列相加),最终得到13+18=31维梯度特征。Then, for the entire target image, it is divided into non-overlapping cells of the same size, and the gradient size and direction of each cell are calculated. DPM preserves the cell units of the HOG map, and then normalizes a cell unit on the image (the 8x8 cell unit in Figure 7) to its four cells in its diagonal neighborhood. Extracting signed HOG gradients, 0-360 degrees will generate 18 gradient vectors, extracting unsigned HOG gradients, 0-180 degrees will generate 9 gradient vectors. DPM only extracts unsigned features, which will generate 4*9=36-dimensional features, and add rows and columns to form 13 feature vectors (as shown in Figure 7, 9 columns are added, and 4 rows are added). In order to further improve the accuracy , the extracted 18-dimensional signed gradient features are also added (18 columns are added as shown in the figure), and finally 13+18=31-dimensional gradient features are obtained.

如图8所示,DPM模型采用了一个8*8分辨率的根滤波器(Root filter)(左)和4*4分辨率的组件滤波器(Part filter)(中)。其中,中图的分辨率为左图的2倍,并且组件滤波器的大小是根滤波器的2倍,因此,看的梯度会更加精细。右图为其高斯滤波后的2倍空间模型。As shown in Figure 8, the DPM model uses an 8*8 resolution Root filter (left) and a 4*4 resolution Part filter (middle). Among them, the resolution of the middle image is 2 times that of the left image, and the size of the component filter is 2 times that of the root filter, so the gradient will be more refined. The picture on the right is the 2x space model after Gaussian filtering.

首先,对输入的图像提取DPM特征图(原始图像的DPM特征图)并进行高斯金字塔上采样(缩放图片),然后再提取高斯金字塔上采样的图片的DPM特征图。将原始图像的DPM特征图和训练好的根滤波器进行卷积运算,从而得到根滤波器的响应图。同时,对于提取到的2倍图像的DPM特征图(高斯金字塔上采样),用训练好的组件滤波器进行卷积运算,得到组件滤波器的响应图。对得到的组件滤波器的响应图进行精细高斯塔下采样操作,如此一来,根滤波器的响应图和组件滤波器的响应图就具有了相同的分辨率。最后,将两者进行加权平均,得到最后的响应图,亮度越大响应效果越好,由此检测到人脸。其中响应值得分公式:First, extract the DPM feature map (DPM feature map of the original image) from the input image and perform Gaussian pyramid upsampling (zooming the image), and then extract the DPM feature map of the Gaussian pyramid upsampled image. Convolve the DPM feature map of the original image and the trained root filter to obtain the response map of the root filter. At the same time, for the extracted DPM feature map (Gaussian pyramid upsampling) of the 2x image, the trained component filter is used for convolution operation to obtain the response map of the component filter. Fine Gausta downsampling is performed on the resulting component filter response map, so that the root filter response map and the component filter response map have the same resolution. Finally, the weighted average of the two is carried out to obtain the final response map. The greater the brightness, the better the response effect, and thus the face is detected. The response value score formula is:

Figure GDA0002899787390000111
Figure GDA0002899787390000111

其中,x0,y0,l0分别为特征点的横坐标、纵坐标、尺度;

Figure GDA0002899787390000112
为根模型的响应分数;
Figure GDA0002899787390000113
为部件模型的响应分;2(x0,y0)表示组件模型的像素为原始的2倍;b为用于与跟模型进行对齐的不同模型组件之间的偏移系数;2(x0,y0)表示组件模型的像素为原始的2倍,所以,像素点*2;vi为像素点和理想检测点之间的偏移系数;其中其部件模型的详细响应得分公式如下:Among them, x 0 , y 0 , and l 0 are the abscissa, ordinate, and scale of the feature point, respectively;
Figure GDA0002899787390000112
is the response score of the root model;
Figure GDA0002899787390000113
is the response score of the component model; 2(x 0 , y 0 ) means that the pixels of the component model are 2 times the original; b is the offset coefficient between different model components used to align with the model; 2(x 0 ,y 0 ) means that the pixel of the component model is twice the original, so, pixel*2; v i is the offset coefficient between the pixel and the ideal detection point; the detailed response score formula of the component model is as follows:

Figure GDA0002899787390000114
Figure GDA0002899787390000114

类似于公式(8),我们希望目标函数(Di,l(x,y))的值越大越好,变量为dx,dy。此外上式中,(x,y)为训练的理想模型的位置;dx,dy是理想模型位置的偏移量,范围是理想位置到图片边缘的位置;Ri,l(x+dx,y+dy)为组件模型的匹配得分;did(dx,dy)为组件的偏移损失得分;di为偏移损失系数;Φd(dx,dy)为组件模型的像素点和组件模型的检测点之间的距离。这个公式表明,组件模型的响应越高,各个组件和其相应的像素点距离越小,则响应分数越高,越有可能是待检测的物体。Similar to formula (8), we hope that the larger the value of the objective function (D i,l (x,y)), the better, the variables are dx, dy. In addition, in the above formula, (x,y) is the position of the ideal model for training; dx, dy is the offset of the ideal model position, and the range is the position from the ideal position to the edge of the picture; R i,l (x+dx,y +dy) is the matching score of the component model; d id (dx, dy) is the offset loss score of the component; d i is the offset loss coefficient; Φ d (dx, dy) is the pixel point sum of the component model The distance between the detection points of the component model. This formula shows that the higher the response of the component model, the smaller the distance between each component and its corresponding pixel point, the higher the response score, and the more likely it is the object to be detected.

训练模型时,要将上面的得出的DPM特征进行训练。DPM在这里用的是Latent-SVM分类法,在这里相对于Linear-SVM分类法增加了Latent变量(潜变量),Latent变量可以用来决定正样本中哪一个样本作为正样本。LSVM中有很多Latent变量,这是因为给定一张正样本的图片,标注完边界框后,需要在某一位置,某一尺度提出一个最大样本作为某一部分的最大样本。图9(a)是一般的Hog+SVM和运用的DPM+Latent–SVM的效果对比图。一般的Hog+SVM和运用的DPM+Latent-SVM的公式如图9(b)所示。When training the model, the DPM features obtained above should be trained. DPM uses the Latent-SVM classification method here. Compared with the Linear-SVM classification method, the Latent variable (latent variable) is added here. The Latent variable can be used to determine which sample in the positive sample is used as a positive sample. There are many Latent variables in LSVM, because given a picture of a positive sample, after marking the bounding box, it is necessary to propose a maximum sample at a certain position and a certain scale as the maximum sample of a certain part. Figure 9(a) is a comparison diagram of the effect of the general Hog+SVM and the applied DPM+Latent-SVM. The general Hog+SVM and the applied DPM+Latent-SVM formula are shown in Figure 9(b).

第三步:人脸特征点定位。在本实施例方案中使用的是LBF算法,采用一个联级的回归器以毫秒为单位进行人脸特征点以及眼部特征点定位。每个回归rt(,)用当前图片I与形状向量

Figure GDA0002899787390000118
来预测更新形状向量,具体公式如下:The third step: facial feature point location. In this embodiment, the LBF algorithm is used, and a cascaded regressor is used to locate the facial feature points and the eye feature points in milliseconds. Each regression r t (,) uses the current image I and shape vector
Figure GDA0002899787390000118
To predict the update shape vector, the specific formula is as follows:

Figure GDA0002899787390000115
Figure GDA0002899787390000115

Figure GDA0002899787390000116
Figure GDA0002899787390000116

Figure GDA0002899787390000117
Figure GDA0002899787390000117

其中

Figure GDA0002899787390000121
表明当前的估计向量S,Xi表示图像I中的脸部特征点的(x,y)坐标。联级中最重要的步骤是回归器rt(,)基于诸如像素灰度特征的预测,这些特征是基于图像I计算和当前形状向量
Figure GDA0002899787390000122
索引出来的。在这个过程中引入了几何形式的不变性,并且随着级联的进行可以人们可以更加确定脸部的精确语义位置正在被索引。in
Figure GDA0002899787390000121
Indicates the current estimated vector S, X i represents the (x, y) coordinates of the facial feature points in the image I. The most important step in the cascade is the regressor r t (,) based on predictions such as pixel grayscale features, which are computed based on the image I and the current shape vector
Figure GDA0002899787390000122
indexed. Invariance of geometric form is introduced in this process, and as the cascade progresses one can be more certain that the precise semantic position of the face is being indexed.

如果初始估计值

Figure GDA0002899787390000123
属于此空间,则确保由集合扩展的输出范围位于训练数据的线性子空间中。这样做不需要对预测实施额外的限制,这极大地简化了本方法。此外,简单地选择初始形状作为根据通用面部检测器的边界框输出居中和缩放的训练数据的平均形状。If the initial estimate
Figure GDA0002899787390000123
belongs to this space, it ensures that the output range extended by the set lies in the linear subspace of the training data. Doing so without imposing additional constraints on the prediction greatly simplifies the method. Furthermore, the initial shape is simply chosen as the average shape of the training data centered and scaled according to the bounding box output of the generic face detector.

接下来就是学习级联中的每个回归器,用训练数据集((I1,S1),......,(In,Sn))来学习回归函数r0,Ii表示一张脸部图片,Si表示一个形状向量。初始化的形状估计和目标更新)

Figure GDA0002899787390000124
如下:The next step is to learn each regressor in the cascade, using the training data set ((I 1 ,S 1 ),...,(I n ,S n )) to learn the regression functions r 0 ,I i represents a face image, and S i represents a shape vector. Initialized shape estimation and target update)
Figure GDA0002899787390000124
as follows:

πi∈{1,......n} (13)π i ∈{1,...n} (13)

Figure GDA00028997873900001212
Figure GDA00028997873900001212

Figure GDA0002899787390000125
Figure GDA0002899787390000125

其中i=(1,......,N)。在这里将这些三元组的总数设置为N=nR,其中R是每个图像I使用的初始化次数。一个图像的每个初始化形状估计是从(S1,......,Sn)统一抽样的,不需要替换。where i=(1,...,N). The total number of these triples is set here as N=nR, where R is the number of initializations each image I uses. Each initialized shape estimate for an image is uniformly sampled from (S 1 ,...,S n ) without replacement.

从这个数据中用伴随平方误差损失的总和的梯度树提升,可以学习到回归函数r0,具体算法如下:From this data, the regression function r 0 can be learned by using the gradient tree boosting with the sum of the squared error loss. The specific algorithm is as follows:

训练数据

Figure GDA0002899787390000126
学习率0<v<1,具体过程是:training data
Figure GDA0002899787390000126
The learning rate is 0<v<1, and the specific process is:

a.初始化:a.Initialization:

Figure GDA0002899787390000127
Figure GDA0002899787390000127

b.对k从1到K:b. From 1 to K for k:

①i=1,…,N:①i=1,…,N:

Figure GDA0002899787390000128
Figure GDA0002899787390000128

②针对回归函数rik拟合一个回归树,给出一个弱回归函数

Figure GDA0002899787390000129
② Fit a regression tree to the regression function rik and give a weak regression function
Figure GDA0002899787390000129

③更新:③Update:

Figure GDA00028997873900001210
Figure GDA00028997873900001210

c.输出:c. Output:

Figure GDA00028997873900001211
Figure GDA00028997873900001211

三元组训练数据进而会更新训练数据为:

Figure GDA0002899787390000131
联级中的下一个回归器r1,被设置如下(t=0):The triplet training data will then update the training data as:
Figure GDA0002899787390000131
The next regressor in the cascade, r 1 , is set as follows (t=0):

Figure GDA0002899787390000132
Figure GDA0002899787390000132

Figure GDA0002899787390000133
Figure GDA0002899787390000133

这个过程被迭代,直到T回归的级联r0,r1,…,rT-1组合后能给予足够的精度水平。This process is iterated until the cascade r 0 ,r 1 ,...,r T-1 combination of T regressions gives a sufficient level of accuracy.

每个回归函数rt的核心是在梯度提升算法中适合残差目标的基于树的回归函数。在回归树的每个分离节点上,我们做出基于两个像素之间强度差异阈值的决定。在基于平均形状定义的坐标系中,测试中用的像素坐标是(u,v)。对于任意形状的脸部图像,我们想要索引具有与其形状相同位置的点就如u和v对于平均形状的点。为了实现这一点,在提取特征之前,可以基于当前的形状估计将图像变形为平均形状。因为我们使用这幅图像的非常稀疏的像素代表,所以更有效的做法是去扭曲这些点的位置而不是去扭曲整个图像。At the heart of each regression function rt is a tree-based regression function that fits the residual objective in the gradient boosting algorithm. At each separation node of the regression tree, we make a decision based on a threshold for the difference in intensity between two pixels. In the coordinate system defined based on the average shape, the pixel coordinates used in the test are (u, v). For an arbitrarily shaped face image, we want to index points that have the same position as their shape as u and v are for points of average shape. To achieve this, the image can be warped to an average shape based on the current shape estimate before extracting features. Since we are using a very sparse pixel representation of this image, it is more efficient to warp the positions of these points rather than warping the entire image.

假设ku是平均形状中与u最接近的面部标志的索引,并定义其与u的偏移为:Suppose k u is the index of the closest facial landmark to u in the mean shape, and define its offset from u as:

Figure GDA0002899787390000134
Figure GDA0002899787390000134

然后,对于图像Ii中定义的形状Si,Ii中的位置这与定义形状图像中的u的定性相似:Then, for a shape S i defined in image I i , position in I i This is qualitatively similar to defining u in a shape image:

Figure GDA0002899787390000135
Figure GDA0002899787390000135

其中si,Ri是比例矩阵和旋转矩阵,这两者被用来最小化平均形状面部标志点和扭曲点之间的平方差之和:where s i , R i are the scale and rotation matrices, both of which are used to minimize the sum of the squared differences between the mean shape facial landmark points and twist points:

Figure GDA0002899787390000136
Figure GDA0002899787390000136

v′被类似地定义。正式的每个分割是涉及3的决定,参数θ=(τ,u,v)并且被应用于每个训练和测试样本。v' is similarly defined. Formally each split is a decision involving 3, the parameter θ = (τ, u, v) and is applied to each training and test sample.

Figure GDA0002899787390000137
Figure GDA0002899787390000137

这里的u'和v′用比例矩阵和旋转矩阵定义。计算相似度转换,在测试时间计算量最大这个过程的一部分,只在每个层次完成一次级联。Here u' and v' are defined by scale and rotation matrices. Computation of similarity transformations, the most computationally intensive part of the process at test time, is done only once at each level of the cascade.

对于每个回归树,我们使用分段常量函数来近似底层函数,其中常量向量适合于每个叶节点。为了训练回归树,我们在每个树节点随机生成一组候选分割,即θ。然后,我们随意地从这些候选者中选择θ,这个做法最小化了平方误差的总和。如果Q是节点上的训练样例的索引集合,则这对应于最小化:For each regression tree, we approximate the underlying function using a piecewise constant function, where a constant vector is fitted to each leaf node. To train a regression tree, we randomly generate a set of candidate splits, θ, at each tree node. We then arbitrarily choose θ from these candidates, which minimizes the sum of squared errors. If Q is the indexed set of training examples on the node, this corresponds to minimizing:

Figure GDA0002899787390000138
Figure GDA0002899787390000138

其中Qθ,S是样本的索引,ri是梯度增强算法中为图像i计算的所有残差的矢量,而μθ,s定义公式如下:where Q θ,S is the index of the sample, ri is the vector of all residuals calculated for image i in the gradient boosting algorithm, and μ θ,s is defined by the following formula:

Figure GDA0002899787390000141
Figure GDA0002899787390000141

最佳优化点可以被很容易地找到,因为如果我们重新排列公式或者忽略掉依赖θ的因素,能看到如下公式关系:The optimal optimization point can be easily found because if we rearrange the formula or ignore the factors that depend on θ, we can see the following formula relationship:

Figure GDA0002899787390000142
Figure GDA0002899787390000142

当评估不同的θ时,我们只需要计算μθ,l正如μθ,r可以通过μ和μθ,l来计算,过程如下:When evaluating different θ, we only need to calculate μ θ,l just as μ θ,r can be calculated by μ and μ θ,l , the process is as follows:

Figure GDA0002899787390000143
Figure GDA0002899787390000143

在每个节点处的决定基于对一对像素处的强度值的差异进行阈值化。这是一个相当简单的测试,但它比单阈值更强大,因为它对全球照明变化的相对不敏感。不幸的是,使用像素差异的缺点是可能的分割(特征)候选的数量在平均图像中的像素数量上是二次的。这使得很难在没有搜索到很多θ的情况下找到好的θ。但是,通过考虑图像数据的结构,这种限制因素在一定程度上可以得到缓解。我们先介绍一个指数The decision at each node is based on thresholding the difference in intensity values at a pair of pixels. This is a fairly simple test, but it is more powerful than single threshold due to its relative insensitivity to changes in global illumination. Unfortunately, the disadvantage of using pixel disparity is that the number of possible segmentation (feature) candidates is quadratic in the number of pixels in the average image. This makes it hard to find a good θ without searching a lot of θ. However, this limitation can be alleviated to some extent by considering the structure of the image data. Let's first introduce an index

p(u,v)αe-λ||u-v|| (29)p(u,v)αe -λ||uv|| (29)

在这个距离范围内的像素分割点容易被选择,这样可以有效降低数据集预测错误的数目。Pixel segmentation points within this distance range are easily selected, which can effectively reduce the number of data set prediction errors.

处理缺失的标签,我们引入了范围在0到1之间的变量wi,j(表示第i个图像的第j个标志点),则得出新的平方差和公式:To deal with missing labels, we introduce a variable w i,j ranging from 0 to 1 (representing the jth marker point of the ith image), then a new sum of squared difference formulas are derived:

Figure GDA0002899787390000144
Figure GDA0002899787390000144

其中Wi是一个向量(wi,i,......wi,p)T变形的对角线矩阵。此外μθ,s的公式如下:where Wi is a vector (wi ,i , ...wi ,p ) T deformed diagonal matrix. In addition, the formula for μ θ,s is as follows:

Figure GDA0002899787390000145
Figure GDA0002899787390000145

梯度增强算法也必须修改以考虑这些权重因子。这可以简单地通过用目标的加权平均值初始化整体模型以及将回归树拟合到加权来完成。此外,拟合回归树的权重值残差算法如下:The gradient boosting algorithm must also be modified to account for these weighting factors. This can be done simply by initializing the overall model with the weighted mean of the targets and fitting a regression tree to the weights. In addition, the weight value residual algorithm for fitting the regression tree is as follows:

Figure GDA0002899787390000146
Figure GDA0002899787390000146

其中,联级迭代效果如图10所示。Among them, the cascade iteration effect is shown in Figure 10.

第四步:人眼特征点定位。本实施例中,使用的模型是基于混合树和共享的部件V池。在这个方法中我们将每个面部地标作为一个部分进行建模,并使用全局混合来捕获由于视点引起的拓扑变化。如11图所示,本实施例中所采用的混合树模型对由于视点引起的拓扑变化进行编码。The fourth step: human eye feature point positioning. In this example, the model used is based on a hybrid tree and a shared V pool of components. In this approach we model each facial landmark as a part and use global blending to capture topological changes due to viewpoints. As shown in FIG. 11, the hybrid tree model adopted in this embodiment encodes the topology changes due to viewpoints.

树结构局部模型:我们写出每个参数线性的树状结构Tm=(Vm,Em),其中,m表明了这个结构是个混合型的,此外

Figure GDA0002899787390000151
我们把一张图片标记为I,并且li=(xi,yi)用来表示位置i处的像素。我们在L部分的得分配置为:Tree-structured local model: We write a linear tree-like structure for each parameter T m = (V m , E m ), where m indicates that the structure is a hybrid, and
Figure GDA0002899787390000151
We label a picture as I, and li = (x i , y i ) is used to denote the pixel at position i . Our scoring configuration for part L is:

S(I,L,m)=Appm(I,L)+Shapem(L)+αm (33)S(I,L, m )=Appm(I,L)+ Shapem (L)+ αm (33)

Figure GDA0002899787390000152
Figure GDA0002899787390000152

式(34)求出了在li位置上i处的模板

Figure GDA0002899787390000153
之和,其中m表示这里是混合型。φ(I,li)表示了在图片I上li像素处的特征向量。Equation (34) finds the template at i at position li
Figure GDA0002899787390000153
The sum, where m indicates that here is a mixed type. φ(I, li ) represents the feature vector at li pixels on image I.

Figure GDA0002899787390000154
Figure GDA0002899787390000154

式(35)表示的是混合类型特定空间L排列的排列得分,其中dx=xi-xj和dy=yi-yj表示第i部分到第j部分的位移。公式中的每个参数(指a,b,c,d)可以被解释为不同部分之间的空间约束。αm表示标量一个偏置的标量。Equation (35) represents the permutation score of the hybrid type specific space L permutation, where dx=x i -x j and dy=y i -y j represent the displacement from the i-th part to the j-th part. Each parameter (referring to a, b, c, d) in the formula can be interpreted as a spatial constraint between different parts. α m denotes a scalar with a bias.

由于本实施例方案中主要运用的是眼部特征点定位的方法,故不考虑整体与部分共享因素。Since the method of locating eye feature points is mainly used in the solution of this embodiment, the whole and part sharing factors are not considered.

求取使式子S(I,L,m)得到最大值的参数L和m的值:Find the values of the parameters L and m that maximize the formula S(I, L, m):

Figure GDA0002899787390000155
Figure GDA0002899787390000155

简单地枚举所有混合型,对于每个混合型找到各部分的最佳配置。Simply enumerate all mixes and find the best configuration for each part for each mix.

因为每个混合型Tm=(Vm,Em)是一个树形的结构,所以内部最大化可以通过动态编程高效完成。由于缺乏空间,可以采用省略消息传递等式的方式。本实施例中的词汇表中不同部分模板的总数是M'|V|,假定每个部分的维度是D并且有N个候选位置。在所有位置评估所有部分的总代价为:Since each hybrid T m = (V m , Em ) is a tree-like structure, internal maximization can be efficiently accomplished by dynamic programming. Due to the lack of space, it is possible to omit the message passing equation. The total number of different part templates in the vocabulary in this embodiment is M'|V|, assuming that the dimension of each part is D and there are N candidate positions. The total cost of evaluating all parts at all locations is:

Figure GDA0002899787390000156
Figure GDA0002899787390000156

之后进行距离转化,信息传递代价就被转化为:O(NM|V|)。这使得本实施例方案的整体模型在部件数量和图像大小方面呈线性。After the distance conversion is performed, the information transfer cost is converted into: O(NM|V|). This makes the overall model of this embodiment scheme linear in the number of parts and the size of the image.

为训练出人眼特征点定位模型。本实施例方案是假定一个全监督场景,在这个场景中有正样本和混合标签以及没有人脸图像的负样本。本实施例方案用结构预测框架有区别地学习形状参数和外观参数。首先需要估计每个混合类型的边缘结构Em。虽然用树结构来得出人类体型模型是很自然的过程,但是对于人眼特征的树形结构还是不太清晰明了的。In order to train the human eye feature point localization model. The solution of this embodiment assumes a fully supervised scene, in which there are positive samples and mixed labels and negative samples without face images. The solution of this embodiment uses the structure prediction framework to learn shape parameters and appearance parameters differently. It is first necessary to estimate the edge structure Em for each mixed type. Although it is a natural process to use a tree structure to derive a human body shape model, the tree structure of human eye features is still not very clear.

实施例方案使用Chow-Liu算法去找出最大相似性树结构,这个最大想实行结构可以最好地解释高斯分布的特征点位置。给定标签的正样本{In,Ln,mn}和负样本{In},实施例方案中将定义一个结构目标预测函数,假定为zn={Ln,mn}。公式S(I,L,m)关于部分模型w,弹性参数(a,b,c,d)以及混合偏置α。把这些参数全部放进一个向量β中,这时我们可以把得分函数写成如下形式:The embodiment solution uses the Chow-Liu algorithm to find the maximum similarity tree structure, which can best explain the feature point positions of the Gaussian distribution. Given positive samples {I n , L n , m n } and negative samples {I n } of labels, a structural target prediction function will be defined in the embodiment solution, assuming z n ={L n ,m n }. The formula S(I,L,m) is about the partial model w, the elastic parameters (a,b,c,d) and the mixing bias α. Putting all these parameters into a vector β, we can write the score function as follows:

S(I,z)=β·Φ(I,z) (38)S(I,z)=β·Φ(I,z) (38)

其中,向量Φ(I,z)是稀疏的,在与混合m相对应的单个区间中具有非零项。where the vector Φ(I,z) is sparse with non-zero entries in a single interval corresponding to the mixture m.

接下来就可以学习到一个如下形式的模型:Next, you can learn a model of the following form:

Figure GDA0002899787390000161
Figure GDA0002899787390000161

式(39)中,C代表目标函数的惩罚项系数(超参数,需要认为调参找出最合适的值),ξn表示对应不同样本的惩罚项(第n个样本的惩罚项),n对应不同的样本,pos和neg分别代表正负样本,K代表目标向量β的个数,k代表对应的目标向量β编号。In formula (39), C represents the penalty item coefficient of the objective function (hyperparameter, it is necessary to adjust the parameters to find the most suitable value), ξ n represents the penalty item corresponding to different samples (the penalty item of the nth sample), n Corresponding to different samples, pos and neg represent positive and negative samples respectively, K represents the number of target vectors β, and k represents the corresponding target vector β number.

如图12所示,为人眼特征点定位结果示意图。As shown in Figure 12, it is a schematic diagram of the positioning result of human eye feature points.

第五步:眨眼检测—计算眼睛的开度并进行疲劳分析。眨眼时眼睛闭合时间变长是驾驶员疲劳的重要标志之一。在前几步的基础上,已经定位到了人眼区域并找到了人眼特征点。为了计算人眼开度,本实施例方案中,首先要排除人眼与摄像头之间的距离的变化,防止其对人眼开度计算造成影响,本实施例方案中对人眼开度采用归一化。在这个基础上,用单位时间内眼睛闭合时间所占比例来判定驾驶员的疲劳程度。具体如下。Step 5: Blink Detection - Calculate the opening of the eyes and perform fatigue analysis. Longer eye closure time when blinking is one of the important signs of driver fatigue. On the basis of the previous steps, the human eye region has been located and the human eye feature points have been found. In order to calculate the opening of the human eye, in the solution of this embodiment, the change of the distance between the human eye and the camera should be excluded first to prevent it from affecting the calculation of the opening of the human eye. unify. On this basis, the degree of fatigue of the driver is determined by the proportion of eye closure time per unit time. details as follows.

首先,根据人眼的开度值,判断人眼的开闭状态。本实施例方案中选取上一步中定位的人眼特征点,找到正对瞳孔处上眼睑与下眼睑处的标定点来计算人眼开度。但是在大量的实验以及经验中发现,人眼与摄像头的距离越远测得的人眼开度越小,人眼与摄像头的距离越近,测得的人眼开度越大。这不利于后期驾驶员疲劳的检测,为此,要归一化处理人眼与摄像头相对位置变化引起的人眼开度异常变化。本实施例方案中利用上一步中眼部特征点定位可以测得人的瞳距l,瞳距与眼睛开度的变化存在线性关系,假设实际测得的人眼开度(相当于上述的人眼开度原始值)为h,归一化后的人眼开度为H,利用下述公式进行人眼开度值修正:First, according to the opening degree value of the human eye, the opening and closing state of the human eye is judged. In this embodiment, the feature points of the human eye located in the previous step are selected, and the calibration points of the upper eyelid and the lower eyelid facing the pupil are found to calculate the opening degree of the human eye. However, it has been found in a large number of experiments and experience that the farther the distance between the human eye and the camera, the smaller the measured human eye opening, and the closer the distance between the human eye and the camera, the greater the measured human eye opening. This is not conducive to the detection of driver fatigue in the later stage. Therefore, it is necessary to normalize the abnormal changes in the opening degree of the human eye caused by the relative position change of the human eye and the camera. In this embodiment, the eye feature point positioning in the previous step can be used to measure the human pupillary distance l, and there is a linear relationship between the pupillary distance and the change of the eye opening. It is assumed that the actually measured human eye opening (equivalent to the above The original value of eye opening) is h, the normalized human eye opening is H, and the following formula is used to correct the human eye opening value:

Figure GDA0002899787390000171
Figure GDA0002899787390000171

其中,C表示一个选择的修正参数。where C represents a selected correction parameter.

接下来是人眼状态划分。本实施例方案中,根据上面获得的人眼开度值,求得人眼的最大张开度,记为MaxW。假设测得的人眼开度为W,其中状态I表示W>80%*MaxWNext is the human eye state division. In the solution of this embodiment, the maximum opening degree of the human eye is obtained according to the value of the opening degree of the human eye obtained above, which is recorded as MaxW. Suppose the measured eye opening is W, where state I means W>80%*MaxW

W>80%MaxW,此时眼睛为完全睁开状态;状态II表示20%*MaxW≤W≤80%*MaxW,眼睛为半睁开状态;状态III表示≤W≤20%*MaxW,眼睛为闭合状态。W>80%MaxW, the eyes are fully open at this time; state II means 20%*MaxW≤W≤80%*MaxW, the eyes are half-open; state III means ≤W≤20%*MaxW, the eyes are closed state.

本实施例中,采用统计该周期(相当于上述的检测周期)内人眼开度小于等于80%最大人眼开度的帧数,记作n,同时统计该周期内人眼开度小于等于20%最大人眼开度的帧数,记作m,即可得到一个比值f,计算公式如下:In this embodiment, the number of frames in which the human eye opening is less than or equal to 80% of the maximum human eye opening in the period (equivalent to the above-mentioned detection period) is counted, and denoted as n, and the human eye opening is less than or equal to the statistics in this period. The frame number of 20% of the maximum human eye opening is recorded as m, and a ratio f can be obtained. The calculation formula is as follows:

Figure GDA0002899787390000172
Figure GDA0002899787390000172

f越接近1,则代表驾驶员越接近疲劳状态。经过大量的实验可以得到一个实验阈值T(相当于上述的疲劳判定阈值),如果f>T则表明该驾驶员处于疲劳状态,以声音形式预警,提醒驾驶员注意疲劳驾驶。The closer f is to 1, the closer the driver is to a fatigued state. After a large number of experiments, an experimental threshold T (equivalent to the above-mentioned fatigue determination threshold) can be obtained. If f>T, it indicates that the driver is in a fatigue state, and an early warning is given in the form of sound to remind the driver to pay attention to fatigue driving.

本实施例方案中,使用DPM算法进行人脸检测,算法的检测准确率大大提升,同时降低误检率与漏检率,提高了光照与人脸姿态的鲁棒性;使用机器学习算法定位人脸特征点与并分别对眼睛特征点进行定位,定位精度非常高同时对光照与姿态有很强的泛化能力,最终算法能够精确估计眼睛开闭程度;对于疲劳检测不仅仅将睁闭眼状态作为主要判据,还包括闭眼时间、单位时间眨眼次数、眼睛开合度等都作为疲劳判据。In the solution of this embodiment, the DPM algorithm is used for face detection, the detection accuracy of the algorithm is greatly improved, and the false detection rate and missed detection rate are reduced at the same time, and the robustness of illumination and face posture is improved; machine learning algorithm is used to locate people The facial feature points and the eye feature points are located separately. The positioning accuracy is very high, and it has a strong generalization ability for illumination and posture. The final algorithm can accurately estimate the degree of eye opening and closing; for fatigue detection, not only the open and closed eyes state As the main criteria, the eye closing time, the number of blinks per unit time, and the degree of eye opening and closing are also used as fatigue criteria.

在一个实施例中,如图13所示,提供了一种驾驶员疲劳方法装置,包括:检测模块1301、处理模块1302、统计模块1303和判别模块1304,其中:In one embodiment, as shown in FIG. 13, a method and apparatus for driver fatigue is provided, including: a detection module 1301, a processing module 1302, a statistics module 1303 and a discrimination module 1304, wherein:

检测模块1301,用于获取目标驾驶员的脸部视频,对所述脸部视频中的各帧脸部图像分别进行人眼开度检测,得到各帧所述脸部图像中的人眼开度值;The detection module 1301 is used to obtain the facial video of the target driver, and perform eye opening detection on each frame of the facial image in the facial video to obtain the eye opening in each frame of the facial image. value;

处理模块1302,用于根据各所述人眼开度值确定第一开度阈值和第二开度阈值,所述第一开度阈值大于所述第二开度阈值;a processing module 1302, configured to determine a first opening threshold and a second opening threshold according to each of the human eye opening values, where the first opening threshold is greater than the second opening threshold;

统计模块1303,用于根据各所述人眼开度值、所述第一开度阈值和所述第二开度阈值,统计所述人眼开度值小于或者等于所述第一开度阈值的第一图像帧数值,以及所述人眼开度值小于或者等于所述第二开度阈值的第二图像帧数值;Statistics module 1303, configured to count the human eye opening value less than or equal to the first opening degree threshold according to each of the human eye opening degree value, the first opening degree threshold value and the second opening degree threshold value The first image frame value of , and the second image frame value of which the eye opening value is less than or equal to the second opening threshold;

判别模块1304,用于若所述第一图像帧数值与所述第二图像帧数值的比值大于预设的疲劳判定阈值,则判定所述目标驾驶员处于疲劳状态。The determination module 1304 is configured to determine that the target driver is in a fatigue state if the ratio of the first image frame value to the second image frame value is greater than a preset fatigue determination threshold.

在其中一个实施例中,检测模块1301可以对各帧所述脸部图像分别进行眼部特征点定位,得到各帧所述脸部图像中的眼部特征点,根据各帧所述脸部图像中的眼部特征点,分别确定各帧所述脸部图像中的人眼瞳距值和人眼开度原始值,根据各帧所述脸部图像中的人眼瞳距值和人眼开度原始值,分别确定各帧所述脸部图像中的人眼开度值。In one embodiment, the detection module 1301 may perform eye feature point positioning on each frame of the face image, to obtain the eye feature points in each frame of the face image, according to the face image of each frame The eye feature points in each frame of the face image respectively determine the value of the interpupillary distance of the human eye and the original value of the human eye opening, according to the value of the interpupillary distance and the human eye opening in the facial image of each frame. The original value of the degree of human eye opening in each frame of the face image is determined respectively.

在一个实施例中,处理模块1302可以根据各所述人眼开度值确定人眼最大张开度值,将所述最大张开度值分别与预设的第一比例系数和第二比例系数相乘,得到所述第一开度阈值和所述第二开度阈值。In one embodiment, the processing module 1302 may determine the maximum opening degree value of the human eye according to each of the human eye opening degree values, and multiply the maximum opening degree value by the preset first proportional coefficient and the second proportional coefficient respectively to obtain the first opening threshold and the second opening threshold.

在其中一个实施例中,检测模块1301可以对各帧所述脸部图像分别进行人脸特征点定位,获得各人脸特征图像;将各所述人脸特征图像分别输入预设的眼睛特征点定位模型,得到各帧所述脸部图像中的眼部特征点;In one embodiment, the detection module 1301 may perform facial feature point positioning on each frame of the facial image to obtain each facial feature image; input each of the facial feature images into preset eye feature points respectively Positioning the model to obtain the eye feature points in the facial image of each frame;

其中,所述眼睛特征点定位模型的训练过程,包括:获取目标图像各个像素点的像素值和各所述像素点出的特征向量;根据所述像素值和所述特征向量,配置树结构局部模型,并确定在L部分的得分函数,所述得分函数为S(I,L,m)=Appm(I,L)+Shapem(L)+αm;通过求取使所述得分函数得到最大值的L和m的值的方式,得到各混合型的各个部分的最佳配置参数;建立训练样本集,所述训练样本集包括设定有标签的正样本和负样本,所述正样本为含有人脸的图像,所述负样本为不含有人脸的图像;根据所述部分模型、所述弹性参数以及所述混合偏置标量构建目标向量,根据所述目标向量修改所述得分函数;根据所述训练样本集、所述最佳配置参数、修改后的所述得分函数以及预先定义的目标预测函数,学习得到所述眼睛特征点定位模型;Wherein, the training process of the eye feature point positioning model includes: acquiring the pixel value of each pixel point of the target image and the feature vector of each pixel point; configuring the local tree structure according to the pixel value and the feature vector model, and determine the score function in the L part, the score function is S(I, L, m)=App m (I, L)+Shape m (L)+α m ; The way to obtain the values of L and m of the maximum value is to obtain the optimal configuration parameters of each part of each hybrid type; a training sample set is established, and the training sample set includes positive samples and negative samples set with labels. The sample is an image containing a human face, and the negative sample is an image that does not contain a human face; a target vector is constructed according to the partial model, the elastic parameter and the mixed bias scalar, and the score is modified according to the target vector function; according to the training sample set, the best configuration parameter, the modified score function and the predefined target prediction function, learn to obtain the eye feature point positioning model;

其中,

Figure GDA0002899787390000181
I表示目标图像,li=(xi,yi)表示所述目标图像的第i个像素点的像素值,w表示部分模型,m表示树结构是混合型,部分模型是指将所述目标图像中的每个面部特征分别作为一个部分进行建模得到,a、b、c和d表示弹性参数,α表示混合偏置标量。in,
Figure GDA0002899787390000181
I represents the target image, li =(x i , y i ) represents the pixel value of the ith pixel of the target image, w represents a partial model, m represents a hybrid tree structure, and a partial model refers to the Each facial feature in the target image is modeled separately as a part, where a, b, c, and d represent the elasticity parameters, and α represents the hybrid bias scalar.

在其中一个实施例中,检测模块1301可以提取第一DPM特征图,所述第一DPM特征图为当前脸部图像的DPM特征图,所述当前脸部图像为任意一帧脸部图像,对所述第一DPM特征图进行采样处理,提取第二DPM特征图,所述第二DPM特征图为对所述第一DPM特征图进行采样处理后的图像的DPM特征图,将所述第一DPM特征图,用预先训练的根滤波器进行卷积运算,得到所述根滤波器的响应图,将N倍的所述第二DPM特征图,用预先训练的组件滤波器进行卷积运算,得到所述组件滤波器的响应图,所述组件滤波器的分辨率为所述根滤波器的分辨率的N倍,N为正整数,根据所述根滤波器的响应图和所述组件滤波器的响应图,得到目标响应图,根据目标响应图获取当前人脸特征图像。In one embodiment, the detection module 1301 may extract a first DPM feature map, where the first DPM feature map is a DPM feature map of a current face image, and the current face image is any frame of face images, The first DPM feature map is sampled to extract a second DPM feature map, where the second DPM feature map is a DPM feature map of an image obtained by sampling the first DPM feature map, and the first DPM feature map is sampled. DPM feature map, perform convolution operation with the pre-trained root filter, obtain the response map of the root filter, and perform convolution operation on N times the second DPM feature map with the pre-trained component filter, Obtain the response diagram of the component filter, the resolution of the component filter is N times the resolution of the root filter, and N is a positive integer, according to the response diagram of the root filter and the component filter The response map of the device is obtained, and the target response map is obtained, and the current face feature image is obtained according to the target response map.

在一个实施例中,提供了一种计算机设备,该计算机设备可以是终端,其内部结构图可以如图14所示。该计算机设备包括通过系统总线连接的处理器、存储器、网络接口、显示屏和输入装置。其中,该计算机设备的处理器用于提供计算和控制能力。该计算机设备的存储器包括非易失性存储介质、内存储器。该非易失性存储介质存储有操作系统和计算机程序。该内存储器为非易失性存储介质中的操作系统和计算机程序的运行提供环境。该计算机设备的网络接口用于与外部的终端通过网络连接通信。该计算机程序被处理器执行时以实现一种人脸特征分析方法。该计算机设备的显示屏可以是液晶显示屏或者电子墨水显示屏,该计算机设备的输入装置可以是显示屏上覆盖的触摸层,也可以是计算机设备外壳上设置的按键、轨迹球或触控板,还可以是外接的键盘、触控板或鼠标等。In one embodiment, a computer device is provided, and the computer device may be a terminal, and its internal structure diagram may be as shown in FIG. 14 . The computer equipment includes a processor, memory, a network interface, a display screen, and an input device connected by a system bus. Among them, the processor of the computer device is used to provide computing and control capabilities. The memory of the computer device includes a non-volatile storage medium, an internal memory. The nonvolatile storage medium stores an operating system and a computer program. The internal memory provides an environment for the execution of the operating system and computer programs in the non-volatile storage medium. The network interface of the computer device is used to communicate with an external terminal through a network connection. The computer program, when executed by the processor, implements a method for analyzing facial features. The display screen of the computer equipment may be a liquid crystal display screen or an electronic ink display screen, and the input device of the computer equipment may be a touch layer covered on the display screen, or a button, a trackball or a touchpad set on the shell of the computer equipment , or an external keyboard, trackpad, or mouse.

本领域技术人员可以理解,图14中示出的结构,仅仅是与本发明方案相关的部分结构的框图,并不构成对本发明方案所应用于其上的计算机设备的限定,具体的计算机设备可以包括比图中所示更多或更少的部件,或者组合某些部件,或者具有不同的部件布置。Those skilled in the art can understand that the structure shown in FIG. 14 is only a block diagram of a part of the structure related to the solution of the present invention, and does not constitute a limitation on the computer equipment to which the solution of the present invention is applied. Include more or fewer components than shown in the figures, or combine certain components, or have a different arrangement of components.

在一个实施例中,提供了一种计算机设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,处理器执行计算机程序时实现以上任意一个实施例中的驾驶员疲劳检测方法。In one embodiment, a computer device is provided, including a memory, a processor, and a computer program stored in the memory and executable on the processor, and the processor implements the driver in any of the above embodiments when the processor executes the computer program Fatigue detection method.

在一个实施例中,提供了一种计算机可读存储介质,其上存储有计算机程序,计算机程序被处理器执行时实现以上任意一个实施例中的驾驶员疲劳检测方法。In one embodiment, a computer-readable storage medium is provided, on which a computer program is stored, and when the computer program is executed by a processor, implements the driver fatigue detection method in any one of the above embodiments.

本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一非易失性计算机可读取存储介质中,该计算机程序在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可包括非易失性和/或易失性存储器。非易失性存储器可包括只读存储器(ROM)、可编程ROM(PROM)、电可编程ROM(EPROM)、电可擦除可编程ROM(EEPROM)或闪存。易失性存储器可包括随机存取存储器(RAM)或者外部高速缓冲存储器。作为说明而非局限,RAM以多种形式可得,诸如静态RAM(SRAM)、动态RAM(DRAM)、同步DRAM(SDRAM)、双数据率SDRAM(DDRSDRAM)、增强型SDRAM(ESDRAM)、同步链路(Synchlink)DRAM(SLDRAM)、存储器总线(Rambus)直接RAM(RDRAM)、直接存储器总线动态RAM(DRDRAM)、以及存储器总线动态RAM(RDRAM)等。Those of ordinary skill in the art can understand that all or part of the processes in the methods of the above embodiments can be implemented by instructing relevant hardware through a computer program, and the computer program can be stored in a non-volatile computer-readable storage In the medium, when the computer program is executed, it may include the processes of the above-mentioned method embodiments. Wherein, any reference to memory, storage, database or other medium used in the various embodiments provided in this application may include non-volatile and/or volatile memory. Nonvolatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory. Volatile memory may include random access memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in various forms such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Road (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.

以上实施例的各技术特征可以进行任意的组合,为使描述简洁,未对上述实施例中的各个技术特征所有可能的组合都进行描述,然而,只要这些技术特征的组合不存在矛盾,都应当认为是本说明书记载的范围。The technical features of the above embodiments can be combined arbitrarily. For the sake of brevity, all possible combinations of the technical features in the above embodiments are not described. However, as long as there is no contradiction in the combination of these technical features, all It is considered to be the range described in this specification.

以上所述实施例仅表达了本发明的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对发明专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本发明构思的前提下,还可以做出若干变形和改进,这些都属于本发明的保护范围。因此,本发明专利的保护范围应以所附权利要求为准。The above-mentioned embodiments only represent several embodiments of the present invention, and the descriptions thereof are specific and detailed, but should not be construed as a limitation on the scope of the invention patent. It should be pointed out that for those of ordinary skill in the art, without departing from the concept of the present invention, several modifications and improvements can also be made, which all belong to the protection scope of the present invention. Therefore, the protection scope of the patent of the present invention should be subject to the appended claims.

Claims (10)

1. A driver fatigue detection method, characterized in that the method comprises:
acquiring a face video of a target driver, and respectively detecting the opening degree of human eyes of each frame of face image in the face video to obtain the opening degree value of human eyes in each frame of face image;
determining a first opening threshold value and a second opening threshold value according to each human eye opening value, wherein the first opening threshold value is larger than the second opening threshold value;
according to each human eye opening value, the first opening threshold and the second opening threshold, counting a first image frame value of which the human eye opening value is smaller than or equal to the first opening threshold and a second image frame value of which the human eye opening value is smaller than or equal to the second opening threshold;
and if the ratio of the first image frame value to the second image frame value is greater than a preset fatigue judgment threshold value, judging that the target driver is in a fatigue state.
2. The method of claim 1, wherein the detecting the eye opening degree of each frame of face image in the face video to obtain the eye opening degree value of each frame of face image comprises:
respectively carrying out eye feature point positioning on each frame of the face image to obtain eye feature points in each frame of the face image;
respectively determining a human eye interpupillary distance value and a human eye opening original value in the face image of each frame according to the eye feature points in the face image of each frame;
and respectively determining the human eye opening value in the face image of each frame according to the human eye interpupillary distance value and the human eye opening original value in the face image of each frame.
3. The driver fatigue detection method according to claim 1 or 2, wherein the determining a first opening degree threshold value and a second opening degree threshold value from each of the human eye opening degree values includes:
determining the maximum eye opening degree value according to the eye opening degree values;
and multiplying the maximum opening degree value by a preset first proportional coefficient and a preset second proportional coefficient respectively to obtain the first opening degree threshold value and the second opening degree threshold value.
4. The method for detecting driver fatigue according to claim 2, wherein the performing eye feature point positioning on the face image of each frame to obtain eye feature points in the face image of each frame includes:
respectively carrying out face feature point positioning on the face image of each frame to obtain each face feature image;
respectively inputting each face feature image into a preset eye feature point positioning model to obtain eye feature points in each frame of the face image;
wherein, the training process of the eye feature point positioning model comprises the following steps:
acquiring a pixel value of each pixel point of a target image and a feature vector of each pixel point;
configuring a tree structure local model according to the pixel values and the feature vectors, and determining a score function in the L part, wherein the score function is S (I, L, m) Appm(I,L)+Shapem(L)+αm
Wherein,
Figure FDA0002899787380000021
i denotes the target image,/i=(xi,yi) Representing the pixel value of the ith pixel point of the target image, w representing a partial model, m representing that the tree structure is a mixed type, the partial model is obtained by modeling each facial feature in the target image as a part, a, b, c and d representing elastic parameters, and alpha representing a mixed offset scalar; phi (I, l)i) Representing l on the target image IiA feature vector at the pixel;
obtaining optimal configuration parameters of each part of each hybrid type by calculating values of L and m which enable the score function to obtain a maximum value;
establishing a training sample set, wherein the training sample set comprises a positive sample and a negative sample which are set with labels, the positive sample is an image containing a human face, and the negative sample is an image not containing the human face;
constructing a target vector according to the partial model, the elasticity parameters and the mixed bias scalar, and modifying the score function according to the target vector;
and learning to obtain the eye characteristic point positioning model according to the training sample set, the optimal configuration parameters, the modified score function and a predefined target prediction function.
5. The method of claim 4, wherein the performing facial feature point location on the facial image of each frame to obtain each facial feature image comprises:
extracting a first DPM feature map, wherein the first DPM feature map is a DPM feature map of a current face image, and the current face image is any one frame of face image; DPM is a target detection algorithm Deformable Part Model;
sampling the first DPM feature map, and extracting a second DPM feature map, wherein the second DPM feature map is a DPM feature map of an image obtained by sampling the first DPM feature map;
performing convolution operation on the first DPM characteristic diagram by using a pre-trained root filter to obtain a response diagram of the root filter;
performing convolution operation on the N times of the second DPM characteristic diagram by using a pre-trained component filter to obtain a response diagram of the component filter, wherein the resolution of the component filter is N times of that of the root filter, and N is a positive integer;
obtaining a target response diagram according to the response diagram of the root filter and the response diagram of the component filter;
and acquiring a current face feature image according to the target response image.
6. The method of detecting driver fatigue as set forth in claim 4, wherein the eye feature point location model is:
Figure FDA0002899787380000031
Figure FDA0002899787380000032
Figure FDA0002899787380000033
Figure FDA0002899787380000034
wherein β represents the target vector, zn={Ln,mnC represents the penalty factor of the objective function, ξnAnd the penalty term of the nth sample is represented, pos and neg respectively represent a positive sample and a negative sample, K represents the number of the target vectors, and K represents the number of the corresponding target vectors.
7. A driver fatigue detecting device, characterized in that the device comprises:
the detection module is used for acquiring a face video of a target driver, and detecting the opening degree of human eyes of each frame of face image in the face video to obtain the opening degree value of human eyes of each frame of face image;
the processing module is used for determining a first opening threshold value and a second opening threshold value according to each human eye opening value, wherein the first opening threshold value is larger than the second opening threshold value;
a counting module, configured to count, according to each of the eye opening values, the first opening threshold and the second opening threshold, a first image frame value of which the eye opening value is smaller than or equal to the first opening threshold, and a second image frame value of which the eye opening value is smaller than or equal to the second opening threshold;
and the judging module is used for judging that the target driver is in a fatigue state if the ratio of the first image frame value to the second image frame value is greater than a preset fatigue judging threshold value.
8. The driver fatigue detection device according to claim 7, characterized in that:
the detection module positions eye feature points of each frame of the face image to obtain eye feature points of each frame of the face image, determines a human eye pupil distance value and a human eye opening original value of each frame of the face image according to the eye feature points of each frame of the face image, and determines a human eye opening value of each frame of the face image according to the human eye pupil distance value and the human eye opening original value of each frame of the face image.
9. A computer device comprising a memory and a processor, the memory storing a computer program, wherein the processor implements the steps of the method of any one of claims 1 to 6 when executing the computer program.
10. A computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, carries out the steps of the method of any one of claims 1 to 6.
CN201811485916.5A 2018-12-06 2018-12-06 Driver fatigue detection method, driver fatigue detection device, computer equipment and storage medium Active CN111291590B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811485916.5A CN111291590B (en) 2018-12-06 2018-12-06 Driver fatigue detection method, driver fatigue detection device, computer equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811485916.5A CN111291590B (en) 2018-12-06 2018-12-06 Driver fatigue detection method, driver fatigue detection device, computer equipment and storage medium

Publications (2)

Publication Number Publication Date
CN111291590A CN111291590A (en) 2020-06-16
CN111291590B true CN111291590B (en) 2021-03-19

Family

ID=71024334

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811485916.5A Active CN111291590B (en) 2018-12-06 2018-12-06 Driver fatigue detection method, driver fatigue detection device, computer equipment and storage medium

Country Status (1)

Country Link
CN (1) CN111291590B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112183220B (en) * 2020-09-04 2024-05-24 广州汽车集团股份有限公司 Driver fatigue detection method and system and computer storage medium thereof
CN112528792B (en) * 2020-12-03 2024-05-31 深圳地平线机器人科技有限公司 Fatigue state detection method, device, medium and electronic equipment
CN116631156A (en) * 2022-02-10 2023-08-22 北京七鑫易维信息技术有限公司 Fatigue driving processing method, device, equipment and storage medium
CN116704482B (en) * 2023-06-09 2025-10-03 阿维塔科技(重庆)有限公司 Fatigue driving detection method, detection system, detection device and computer-readable storage medium
CN119229422B (en) * 2024-12-03 2025-02-18 福建警察学院 A method and system for detecting fatigue driving under low illumination

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7742621B2 (en) * 2006-06-13 2010-06-22 Delphi Technologies, Inc. Dynamic eye tracking system
CN100462047C (en) * 2007-03-21 2009-02-18 汤一平 Safe driving auxiliary device based on omnidirectional computer vision
CN101732055B (en) * 2009-02-11 2012-04-18 北京智安邦科技有限公司 driver fatigue detection method and system
CN102013013B (en) * 2010-12-23 2013-07-03 华南理工大学广州汽车学院 Fatigue driving monitoring method
CN104013414B (en) * 2014-04-30 2015-12-30 深圳佑驾创新科技有限公司 A kind of Study in Driver Fatigue State Surveillance System based on intelligent movable mobile phone
CN104881955B (en) * 2015-06-16 2017-07-18 华中科技大学 A kind of driver tired driving detection method and system
CN105354985B (en) * 2015-11-04 2018-01-12 中国科学院上海高等研究院 Fatigue driving monitoring apparatus and method
CN105574487A (en) * 2015-11-26 2016-05-11 中国第一汽车股份有限公司 Facial feature based driver attention state detection method
US10262219B2 (en) * 2016-04-21 2019-04-16 Hyundai Motor Company Apparatus and method to determine drowsiness of a driver
CN106530623B (en) * 2016-12-30 2019-06-07 南京理工大学 A kind of fatigue driving detection device and detection method
CN107169437A (en) * 2017-05-11 2017-09-15 南宁市正祥科技有限公司 The method for detecting fatigue driving of view-based access control model

Also Published As

Publication number Publication date
CN111291590A (en) 2020-06-16

Similar Documents

Publication Publication Date Title
CN111291590B (en) Driver fatigue detection method, driver fatigue detection device, computer equipment and storage medium
CN111709409B (en) Face living body detection method, device, equipment and medium
US12099577B2 (en) Object recognition method and apparatus, electronic device, and readable storage medium
Lu et al. Dense and sparse reconstruction error based saliency descriptor
Deng et al. M3 csr: Multi-view, multi-scale and multi-component cascade shape regression
CN115050064B (en) Human face liveness detection method, device, equipment and medium
Kim et al. Color–texture segmentation using unsupervised graph cuts
Karavasilis et al. Visual tracking using the Earth Mover's Distance between Gaussian mixtures and Kalman filtering
CN107330412B (en) A face age estimation method based on deep sparse representation
CN111291607B (en) Driver distraction detection method, device, computer equipment and storage medium
CN101996308A (en) Human face identification method and system and human face model training method and system
CN112651996B (en) Target detection tracking method, device, electronic equipment and storage medium
CN113642520B (en) Double-task pedestrian detection method with head information
CN109993021A (en) Face detection method, device and electronic device
Efraty et al. Facial component-landmark detection
CN110175500B (en) Finger vein comparison method, device, computer equipment and storage medium
CN105701469A (en) Robust population counting method based on cost-sensitive sparse linear regression
Lee et al. Learning non-homogenous textures and the unlearning problem with application to drusen detection in retinal images
Rosin Measuring sigmoidality
CN115880483A (en) Neural network model distillation method, device and storage medium
CN115797990A (en) Image classification method, image processing method, image classification device and storage medium
Liu et al. An adaptive feature-fusion method for object matching over non-overlapped scenes
CN114565940B (en) A target detection method and device
Reddy et al. Enhancement of accuracy in face recognition system using K-Nearest Neighbour algorithm over support vector machine
Asefaw Re-recognition of vehicles for enhanced insights on road traffic

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant