CN111899193B - Criminal investigation photographing system and method based on low-illumination image enhancement algorithm - Google Patents
Criminal investigation photographing system and method based on low-illumination image enhancement algorithm Download PDFInfo
- Publication number
- CN111899193B CN111899193B CN202010753913.6A CN202010753913A CN111899193B CN 111899193 B CN111899193 B CN 111899193B CN 202010753913 A CN202010753913 A CN 202010753913A CN 111899193 B CN111899193 B CN 111899193B
- Authority
- CN
- China
- Prior art keywords
- enhancement
- image
- low
- network
- recognition
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30168—Image quality inspection
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Probability & Statistics with Applications (AREA)
- Multimedia (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
Abstract
本发明属于图像增强技术领域,公开了一种基于低照度图像增强算法的刑侦摄影系统及方法,摄影端利用摄影设备进行图像数据获取与对象捕捉,并将获取的图像数据传送至增强端;增强端利用构建的基于增强网络模块生成对抗网络的低照度图像增强模型对传输的低照度图像进行增强,并运用对抗网络算法进行图像识别,为未进行过识别的种类贴上标签;同时提取存储后台中已完成识别的类似标签进行辅助识别;将完成识别增强的图片传输至摄影端以及存储后台分别进行输出与存储。本发明低照度图像增强效果最好、增强效率高、应成体系的系统可应用于各种复杂场景、选取了最适用于刑侦摄影的算法技术。
The present invention belongs to the field of image enhancement technology, and discloses a criminal investigation photography system and method based on a low-light image enhancement algorithm. The photography end uses a photography device to acquire image data and capture objects, and transmits the acquired image data to the enhancement end; the enhancement end uses a low-light image enhancement model based on an enhancement network module to generate an adversarial network to enhance the transmitted low-light image, and uses an adversarial network algorithm to perform image recognition, and labels the types that have not been recognized; at the same time, similar labels that have been recognized in the storage background are extracted for auxiliary recognition; the images that have completed recognition and enhancement are transmitted to the photography end and the storage background for output and storage respectively. The present invention has the best low-light image enhancement effect, high enhancement efficiency, and a system that can be applied to various complex scenes, and selects the algorithm technology that is most suitable for criminal investigation photography.
Description
技术领域Technical Field
本发明属于图像增强技术领域,尤其涉及一种基于低照度图像增强算法的刑侦摄影系统及方法。The present invention belongs to the technical field of image enhancement, and in particular relates to a criminal investigation photography system and method based on a low-illumination image enhancement algorithm.
背景技术Background technique
目前,作为刑事侦查中的一种重要的途径,刑侦摄影技术的掌握对刑事案件线索的掌握十分重要。刑侦摄影技术人员需要使用专业的器材准确,客观的将实际情况记录下来,这些图像为案件的告破提供科学的图像和证据。同时大名鼎鼎的“天网监控系统”也将摄影与识别技术相结合,进而完成相关刑事任务。At present, as an important way in criminal investigation, the mastery of criminal investigation photography technology is very important for grasping clues of criminal cases. Criminal investigation photography technicians need to use professional equipment to accurately and objectively record the actual situation. These images provide scientific images and evidence for solving the case. At the same time, the famous "Skynet Monitoring System" also combines photography with recognition technology to complete related criminal tasks.
现有技术1比较早的用深度学习方法完成低光照增强任务的文章,证明了基于合成数据训练的堆叠稀疏去噪自编码器能够对的低光照有噪声图像进行增强和去噪。模型训练基于图像块(patch),采用sparsity regularized reconstruction loss作为损失函数,主要贡献如下:(1)提出了一种训练数据生成方法(即伽马校正和添加高斯噪声)来模拟低光环境。(2)探索了两种类型的网络结构:(a)LLNet,同时学习对比度增强和去噪;(b)S-LLNet,使用两个模块分阶段执行对比度增强和去噪。(3)在真实拍摄到的低光照图像上进行了实验,证明了用合成数据训练的模型的有效性。(4)可视化了网络权值,提供了关于学习到的特征的insights。Prior Art 1 is an earlier article that uses deep learning methods to complete low-light enhancement tasks. It proves that stacked sparse denoising autoencoders trained on synthetic data can enhance and denoise low-light noisy images. The model training is based on image patches and uses sparsity regularized reconstruction loss as the loss function. The main contributions are as follows: (1) A training data generation method (i.e., gamma correction and adding Gaussian noise) is proposed to simulate low-light environments. (2) Two types of network structures are explored: (a) LLNet, which learns contrast enhancement and denoising simultaneously; (b) S-LLNet, which uses two modules to perform contrast enhancement and denoising in stages. (3) Experiments are conducted on real low-light images to prove the effectiveness of the model trained with synthetic data. (4) The network weights are visualized to provide insights about the learned features.
第一种利用LLNet进行两种操作的方法中,自动编码器模块由多层隐藏单元组成,其中编码器通过无监督学习进行训练,解码器权重从编码器移置,随后通过反向传播进行误差微调In the first approach, which uses LLNet for both operations, the autoencoder module consists of multiple layers of hidden units, where the encoder is trained via unsupervised learning and the decoder weights are transferred from the encoder, followed by error fine-tuning via back-propagation.
在使用两个模块分阶段进行对比实验的方法b中,具有同时增强对比度和降噪模块的LLNet,也具有顺序对比增强和降噪模块的S-LLNet。In method b, the comparative experiments are conducted in stages using two modules, LLNet with simultaneous contrast enhancement and denoising modules, and S-LLNet with sequential contrast enhancement and denoising modules.
现有技术2引入了CNN,传统的multi-scale Retinex(MSR)方法可以看作是有着不同高斯卷积核的前馈卷积神经网络,并进行了详细论证。接着,仿照MSR的流程,他们提出了MSR-net,直接学习暗图像到亮图像的端到端映射。训练数据采用的是用PS调整过的高质量图像和对应的合成低光照图像(随机减少亮度、对比度,伽马校正)。损失函数为带正则项的误差矩阵的F-范数平方,即误差平方和。MSR-net包括三个模块:多尺度对数变换,卷积差分,颜色恢复。Prior art 2 introduced CNN. The traditional multi-scale Retinex (MSR) method can be regarded as a feed-forward convolutional neural network with different Gaussian convolution kernels, and a detailed demonstration was given. Then, following the process of MSR, they proposed MSR-net, which directly learns the end-to-end mapping of dark images to bright images. The training data used high-quality images adjusted with PS and corresponding synthetic low-light images (randomly reducing brightness and contrast, and gamma correction). The loss function is the F-norm square of the error matrix with regularization, that is, the sum of squared errors. MSR-net includes three modules: multi-scale logarithmic transformation, convolution difference, and color restoration.
现有技术3其实主要关注单图像对比度增强(SICE),针对的是欠曝光和过曝光情形下的低对比度问题。其主要贡献如下:(1)构建了一个多曝光图像数据集,包括了不同曝光度的低对比度图像以及对应的高质量参考图像。(2)提出了一个两阶段的增强模型。第一阶段先用加权最小二乘(WLE)滤波方法将原图像分解为低频成分和高频成分,然后对两种成分分别进行增强;第二阶段对增强后的低频和高频成分融合,然后再次增强,输出结果。由于单阶段CNN的增强结果并不令人满意,且存在色偏现象,这可能是因为单阶段CNN难以平衡图像的平滑成分与纹理成分的增强效果,故设计成两阶段网络结构,其中模型第一阶段的Decomposition步骤采用的是传统方法,而后面介绍的Retinex-Net使用CNN实现了。The existing technology 3 actually focuses on single image contrast enhancement (SICE), which targets the low contrast problem under underexposure and overexposure. Its main contributions are as follows: (1) A multi-exposure image dataset is constructed, including low-contrast images of different exposures and corresponding high-quality reference images. (2) A two-stage enhancement model is proposed. In the first stage, the weighted least squares (WLE) filtering method is used to decompose the original image into low-frequency components and high-frequency components, and then the two components are enhanced separately; in the second stage, the enhanced low-frequency and high-frequency components are fused, and then enhanced again to output the result. Since the enhancement result of the single-stage CNN is not satisfactory and there is a color cast phenomenon, this may be because the single-stage CNN is difficult to balance the enhancement effect of the smooth component and the texture component of the image, so it is designed into a two-stage network structure, in which the Decomposition step of the first stage of the model adopts the traditional method, and the Retinex-Net introduced later is implemented using CNN.
随着图像识别技术的发展,以人脸,身形等因素作为身份识别的标签已经被广泛应用。人脸识别等技术给本发明的日常生活提供了很多便利,在刑侦方面,通过人脸,身形等因素作为识别的特征识别罪犯身份,以及利用物品识别来搜索人眼容易忽略的细节,进而在刑侦摄影领域引入了图像识别技术。作为刑事侦查中的一种重要的途径,刑侦摄影技术的掌握对刑事案件线索的掌握十分重要。刑侦摄影技术人员需要使用专业的器材准确,客观的将实际情况记录下来,这些图像为案件的告破提供科学的图像和证据。同时大名鼎鼎的“天网监控系统”也将摄影与识别技术相结合,进而完成相关刑事任务。With the development of image recognition technology, face, body shape and other factors have been widely used as labels for identity recognition. Face recognition and other technologies provide a lot of convenience to the daily life of the present invention. In criminal investigation, the identity of the criminal is identified by face, body shape and other factors as identification features, and object identification is used to search for details that are easily overlooked by the human eye, thereby introducing image recognition technology in the field of criminal investigation photography. As an important way in criminal investigation, the mastery of criminal investigation photography technology is very important for grasping clues in criminal cases. Criminal investigation photography technicians need to use professional equipment to accurately and objectively record the actual situation. These images provide scientific images and evidence for solving the case. At the same time, the famous "Skynet Monitoring System" also combines photography with recognition technology to complete related criminal tasks.
然而,显示场景的复杂性,对刑侦摄影与图像识别技术的结合提出了巨大的“挑战”。刑侦的现场可能出现在各种场所,场所多样性所带来的各种噪声,提高了高精度的图像生成的难度。因此,增强对于存在噪声的图片的分析是当今研究的重点因为光影对于图片的影响范围最为广泛,且对图片质量的影响也最大,并且刑侦领域中的犯罪现场更多的出现在人迹罕至,阴暗的场所,并且出现在夜晚的的比例远比出现在白天的情况多很多。但在现行研究结果中,对于上述问题的处理并没有十分完善的手段。因此,研究低照度图像的增强与刑侦摄影相结合的方式具有重要的理论意义和实际应用价值。However, the complexity of the display scene poses a huge "challenge" to the combination of criminal investigation photography and image recognition technology. The scene of criminal investigation may appear in various places, and the various noises brought by the diversity of places increase the difficulty of generating high-precision images. Therefore, enhancing the analysis of pictures with noise is the focus of current research because light and shadow have the widest impact on pictures and the greatest impact on picture quality. In addition, crime scenes in the field of criminal investigation appear more in places with few people and dark places, and the proportion of scenes appearing at night is much higher than that appearing during the day. However, in the current research results, there is no perfect means to deal with the above problems. Therefore, it is of great theoretical significance and practical application value to study the combination of low-light image enhancement and criminal investigation photography.
上述方法虽然都部分实现了对于低照度图像的增强,但是在后续的实验环节的对比中发现,其实验效果远不佳。Although the above methods have partially achieved the enhancement of low-light images, it was found in the comparison of subsequent experimental links that their experimental results were far from satisfactory.
通过上述分析,现有技术存在的问题及缺陷为:Through the above analysis, the problems and defects of the prior art are as follows:
(1)现有低照度图像增强算法没有解决各种复杂场景产生的各个种类的噪声对于生成图片质量的影响。(1) Existing low-light image enhancement algorithms do not address the impact of various types of noise generated by various complex scenes on the quality of generated images.
(2)现有低照度图像增强算法无法有效解决光影问题。(2) Existing low-light image enhancement algorithms cannot effectively solve light and shadow problems.
(3)现有技术缺乏成体系的建模方法,无法建立可以适用于各种情景的低照度图片增强系统。(3) The existing technology lacks a systematic modeling method and cannot establish a low-light image enhancement system that can be applied to various scenarios.
解决以上问题及缺陷的难度为:The difficulty of solving the above problems and defects is:
(1)复杂的场景以及多种类噪声的存在对于增强模块的鲁棒性(适应性)提出很高的要求。(1) Complex scenes and the presence of multiple types of noise place high demands on the robustness (adaptability) of the enhancement module.
(2)光影问题的存在很难通过传统的亮度增强方式来解决。(2) The existence of light and shadow problems is difficult to solve through traditional brightness enhancement methods.
(3)成体系的建模方法对于增强算法的适应性提出了很高的要求。(3) Systematic modeling methods place high demands on enhancing the adaptability of algorithms.
解决以上问题及缺陷的意义为:The significance of solving the above problems and defects is:
(1)刑侦摄影所拍摄对象的场景存在复杂多样的情况,解决上文中所提第一类缺陷可以很好的帮助刑侦摄影技术适应各类型的环境。(1) The scenes captured by criminal investigation photography are complex and diverse. Solving the first type of defects mentioned above can help criminal investigation photography technology adapt to various types of environments.
(2)刑侦摄影所应用的多数场景如阴暗的拐角,地下停车场等此类型的场景光影方面的问题普遍存在。解决上述第二类缺陷可以帮助刑侦摄影在此类场景有效的发挥作用。(2) Most scenes used in criminal investigation photography, such as dark corners and underground parking lots, have common lighting and shadow problems. Solving the second type of defect can help criminal investigation photography play an effective role in such scenes.
(3)形成成体系的建模方式可以帮助刑侦摄影技术有效的对接各类型应用。(3) A systematic modeling approach can help criminal investigation photography technology effectively connect to various types of applications.
发明内容Summary of the invention
针对现有技术存在的问题,本发明提供了一种基于低照度图像增强算法的刑侦摄影系统。In view of the problems existing in the prior art, the present invention provides a criminal investigation photography system based on a low-illumination image enhancement algorithm.
本发明是这样实现的,一种基于低照度图像增强算法的刑侦摄影系统,所述基于低照度图像增强算法的刑侦摄影系统包括:摄影端,增强端与储存后台;The present invention is implemented as follows: a criminal investigation photography system based on a low-light image enhancement algorithm, the criminal investigation photography system based on a low-light image enhancement algorithm comprises: a photography end, an enhancement end and a storage background;
所述摄影端,与增强端连接;包括摄影模块与输出模块;所述摄影模块,用于进行图像数据获取与对象捕捉,并将获取的图像数据传送至增强端;所述输出模块用于输出增强后的图像;The photographing end is connected to the enhancing end; it includes a photographing module and an output module; the photographing module is used to acquire image data and capture objects, and transmit the acquired image data to the enhancing end; the output module is used to output the enhanced image;
所述增强端,与摄影端以及储存后台连接,包括增强模块,识别模块与传输模块;The enhancement end is connected to the photographing end and the storage background, and includes an enhancement module, a recognition module and a transmission module;
所述增强模块,用于利用基于增强网络模块生成对抗网络的低照度图像增强模型对传输的低照度图像进行增强;所述识别模块,用于运用对抗网络算法进行图像识别;同时用于提取后台中已完成识别的类似标签进行辅助识别,也可给未进行过识别的种类贴上标签;所述传输模块,用于将完成识别增强的图片传输至摄影端中的输出模块,同时用于将增强、识别结果与储存后台中未进行过存储的标签传输至后台进行保存;The enhancement module is used to enhance the transmitted low-light image by using a low-light image enhancement model based on the enhancement network module to generate an adversarial network; the recognition module is used to perform image recognition using an adversarial network algorithm; it is also used to extract similar tags that have been recognized in the background for auxiliary recognition, and it can also label types that have not been recognized; the transmission module is used to transmit the image that has completed recognition and enhancement to the output module in the camera end, and is also used to transmit the enhancement and recognition results and tags that have not been stored in the storage background to the background for storage;
所述存储后台,与增强端连接,用于保存增强、识别结果与识别种类标签。The storage backend is connected to the enhancement end and is used to store enhancement, recognition results and recognition category labels.
本发明的另一目的在于提供一种应用于所述基于低照度图像增强算法的刑侦摄影系统的基于低照度图像增强算法的刑侦摄影方法,所述基于低照度图像增强算法的刑侦摄影方法包括:Another object of the present invention is to provide a criminal investigation photography method based on a low illumination image enhancement algorithm applied to the criminal investigation photography system based on the low illumination image enhancement algorithm, the criminal investigation photography method based on the low illumination image enhancement algorithm comprising:
步骤一,摄影端利用摄影设备进行图像数据获取与对象捕捉,并将获取的图像数据传送至增强端;Step 1: The photographing end uses a photographing device to acquire image data and capture an object, and transmits the acquired image data to the enhancing end;
步骤二,增强端利用构建的基于增强网络模块生成对抗网络的低照度图像增强模型对传输的低照度图像进行增强,并运用对抗网络算法进行图像识别,为未进行过识别的种类贴上标签;同时提取存储后台中已完成识别的类似标签进行辅助识别;Step 2: The enhancement end uses the low-light image enhancement model based on the enhancement network module to generate an adversarial network to enhance the transmitted low-light image, and uses the adversarial network algorithm to perform image recognition and label the types that have not been recognized; at the same time, similar labels that have been recognized in the storage background are extracted for auxiliary recognition;
步骤三,将完成识别增强的图片传输至摄影端以及存储后台分别进行输出与存储。Step three: transmit the enhanced image to the camera end and storage background for output and storage respectively.
进一步,步骤一中,会根据摄影端硬件设备的不同,如不同相机的不同的特点,以及所需求捕捉的对象不同,进行硬件设备功能的特殊设置。其捕捉与传输技术已广泛的应用于各种类型的智能设备。Furthermore, in step 1, special settings of the hardware device functions will be made according to the different hardware devices at the photography end, such as the different characteristics of different cameras and the different objects to be captured. Its capture and transmission technology has been widely used in various types of smart devices.
其捕捉技术大体流程如下:The capture technology generally follows the following process:
生成camera controller(各类型的摄影控制系统)作为操作引擎接口。Generate camera controller (various types of photography control systems) as the operation engine interface.
初始化摄影设备引擎。Initialize the camera device engine.
进行异步操开始掌控相机整体。Perform asynchronous operations to start controlling the entire camera.
异步操作启动相机,同时调用相机观察系统。The asynchronous operation starts the camera and calls the camera observation system.
利用相机观察系统开始准备图像捕捉,同时设置引擎为空闲状态。Start preparing for image capture using the camera observation system and set the engine to idle state.
需要拍摄图片后,调用相机控制系统,执行引擎中的工程命令,之后进行操作。When you need to take a picture, call the camera control system, execute the engineering commands in the engine, and then perform the operation.
进行捕捉图片异步操作,完成后调用相机引擎中的图片捕捉完成指令完成操作。Perform asynchronous image capture operations. After completion, call the image capture completion instruction in the camera engine to complete the operation.
进一步,步骤三中,图像传输已经广泛的运用于各种类型的智能设备。在图像进行传输,并且完成储存这个过程中,本发明需要完成图像数字化处理,凭借此来存储图像的数字数据。Furthermore, in step 3, image transmission has been widely used in various types of smart devices. In the process of image transmission and storage, the present invention needs to complete image digitization processing to store the digital data of the image.
其数字化处理大体流程如下:The general process of digital processing is as follows:
采样:采样的实质就是要用多少点来描述一幅图像,采样结果质量的高低就是用前面所说的图像分辨率来衡量.Sampling: The essence of sampling is how many points are used to describe an image. The quality of the sampling result is measured by the image resolution mentioned above.
量化:量化是指要使用多大范围的数值来表示图像采样之后的每一个点.量化的结果是图像能够容纳的颜色总数,它反映了采样的质量.Quantization: Quantization refers to the range of values used to represent each point after image sampling. The result of quantization is the total number of colors that the image can accommodate, which reflects the quality of sampling.
压缩编码:数字化后得到的图像数据量十分巨大,必须采用编码技术来压缩其信息量。在一定意义上讲,编码压缩技术是实现图像传输与储存的关键。Compression coding: The amount of image data obtained after digitization is very huge, and coding technology must be used to compress its information. In a sense, coding compression technology is the key to achieving image transmission and storage.
现有技术一般的压缩编码方法都是混合编码,即将变换编码,运动估计和运动补偿,以及熵编码三种方式相结合来进行压缩编码。The general compression coding method in the prior art is hybrid coding, that is, combining transform coding, motion estimation and motion compensation, and entropy coding to perform compression coding.
变换编码:消去除图像的帧内冗余。Transform coding: Eliminate intra-frame redundancy of images.
运动估计和运动补偿:去除图像的帧间冗余。Motion estimation and motion compensation: remove inter-frame redundancy in images.
熵编码:进一步提高压缩的效率。Entropy coding: further improves compression efficiency.
1)变换编码1) Transform coding
变换编码的作用是将空间域描述的图像信号变换到频率域,然后对变换后的系数进行编码处理。一般来说,图像在空间上具有较强的相关性,变换到频率域可以实现去相关和能量集中。The function of transform coding is to transform the image signal described in the spatial domain into the frequency domain, and then encode the transformed coefficients. Generally speaking, images have strong spatial correlation, and transforming them into the frequency domain can achieve decorrelation and energy concentration.
2)熵编码2) Entropy Coding
熵编码是因编码后的平均码长接近信源熵值而得名。熵编码多用可变字长编码(VLC,Variable Length Coding)实现。其基本原理是对信源中出现概率大的符号赋予短码,对于出现概率小的符号赋予长码,从而在统计上获得较短的平均码长。Entropy coding is named because the average code length after coding is close to the entropy value of the source. Entropy coding is often implemented using variable length coding (VLC). Its basic principle is to assign short codes to symbols with a high probability of appearing in the source, and assign long codes to symbols with a low probability of appearing, thereby obtaining a shorter average code length statistically.
3)运动估计和运动补偿3) Motion Estimation and Motion Compensation
运动估计(Motion Estimation)和运动补偿(Motion Compensation)是消除图像序列时间方向相关性的有效手段。Motion estimation and motion compensation are effective means to eliminate the temporal correlation of image sequences.
运动估计技术一般将当前的输入图像分割成若干彼此不相重叠的小图像子块,例如一帧图像的大小为1280x720,首先将其以网格状的形式分成40x45个尺寸为16x16的彼此没有重叠的图像块,然后在前一图像或者后一个图像某个搜索窗口的范围内为每一个图像块寻找一个与之最为相似的图像块。这个搜寻的过程叫做运动估计。通过计算最相似的图像块与该图像块之间的位置信息,可以得到一个运动矢量。这样在编码过程中就可以将当前图像中的块与参考图像运动矢量所指向的最相似的图像块相减,得到一个残差图像块,由于残差图像块中的每个像素值很小,所以在压缩编码中可以获得更高的压缩比。这个相减过程叫运动补偿。Motion estimation technology generally divides the current input image into several small non-overlapping image sub-blocks. For example, if the size of a frame image is 1280x720, it is first divided into 40x45 non-overlapping image blocks of size 16x16 in a grid-like form, and then a most similar image block is found for each image block within a certain search window of the previous image or the next image. This search process is called motion estimation. By calculating the position information between the most similar image block and the image block, a motion vector can be obtained. In this way, during the encoding process, the block in the current image can be subtracted from the most similar image block pointed to by the motion vector of the reference image to obtain a residual image block. Since each pixel value in the residual image block is very small, a higher compression ratio can be obtained in compression coding. This subtraction process is called motion compensation.
进一步,步骤二中,所述基于增强网络模块生成对抗网络的低照度图像增强模型构建方法包括:Further, in step 2, the method for constructing a low-light image enhancement model based on an enhancement network module generating an adversarial network includes:
(1)获取高质量的图像对作为训练数据集训练生成器网络G;(1) Obtain high-quality image pairs as training datasets to train the generator network G;
(2)从训练数据集随机采样m个低照度图片对其中,m表示训练批次的大小,Ix表示低照度图片,Iy表示真实照度图片,Iadv表示判别器的输入;(2) Randomly sample m low-light image pairs from the training dataset Where m represents the size of the training batch, I x represents the low-light image, I y represents the real-light image, and I adv represents the input of the discriminator;
(3)固定判别网络的输入为Iadv={0,0,…,0},长度为m;(3) The input of the discriminant network is fixed to I adv = {0,0,…,0}, with a length of m;
(4)最小化生成器网络总体损失:(4) Minimize the overall loss of the generator network:
Lossgen=ωaLa+ωadvLadv+ωconLcon+ωtvLtv+ωcolLcol;Loss gen =ω a L a +ω adv L adv +ω con L con +ω tv L tv +ω col L col ;
(5)训练判别器网络,随机初始化判别网络的输入为Iadv={1,0,…,0},长度为m;(5) Train the discriminator network and randomly initialize the input of the discriminator network to I adv = {1, 0, …, 0}, with a length of m;
(6)从训练数据集随机采样m个低照度图片 (6) Randomly sample m low-light images from the training dataset
(7)最大化判别器网络总体损失: (7) Maximize the overall loss of the discriminator network:
进一步,所述基于增强网络模块生成对抗网络的低照度图像增强模型包括:Furthermore, the low-light image enhancement model based on the enhanced network module generating the adversarial network includes:
所述低照度图像增强模型包括添加有增强网络的生成器、判别器以及损失函数;The low-light image enhancement model includes a generator, a discriminator and a loss function added with an enhancement network;
所述添加有增强网络的生成器,基于完全卷积网络,由多个残差块和卷积块2部分组成;用于将输入图像作为一个整体转换成一个在新空间中类似的图片;The generator with the enhanced network added is based on a fully convolutional network and consists of two parts: a plurality of residual blocks and a convolutional block; it is used to convert the input image as a whole into a similar picture in a new space;
所述判别器,用于同时接收生成器生成的图片和真实图片,产生真假的预测值;The discriminator is used to simultaneously receive the image generated by the generator and the real image, and generate a true or false prediction value;
所述损失函数为:Loss=ωaLa+ωadvLadv+ωconLcon+ωtvLtv+ωcolLcol;The loss function is: Loss = ω a L a + ω adv L adv + ω con L con + ω tv L tv + ω col L col ;
其中La,Ladv,Lcon,Ltv,Lcol分别表示注意力损失,对抗损失,内容损失,总变差损失,颜色损失,ωa,ωadv,ωcon,ωtv,ωcol分别表示其损失对应权重。Among them, La , Ladv , Lcon , Ltv , Lcol represent attention loss, adversarial loss, content loss, total variation loss, and color loss respectively, and ωa , ωadv , ωcon , ωtv , ωcol represent the corresponding weights of their losses respectively.
进一步,所述增强网络模块包括:Further, the enhanced network module includes:
所述增强网络模块,所述增强网络模块包含2层3x3卷积层以及特征变换层;用于克服图片低对比度的缺点并改善细节,进行图片效果增强;The enhanced network module includes two 3x3 convolutional layers and a feature transformation layer; it is used to overcome the disadvantage of low contrast of the image and improve the details to enhance the image effect;
所述3x3卷积层第一层用于进行特征提取,实现从rgb通道到多个特征;The first layer of the 3x3 convolutional layer is used for feature extraction, realizing the conversion from RGB channels to multiple features;
所述卷积层第一层之后为特征变换层,所述特征变换层为残差模块;所述特征变换层,用于通过连接多个残差单元进行复杂的特征变换;The first convolutional layer is followed by a feature transformation layer, which is a residual module; the feature transformation layer is used to perform complex feature transformation by connecting multiple residual units;
所述特征变换层后为两层卷积层,所述卷积层用于恢复rgb图片,实现多特征转换为rgb图片;所述每次卷积之后进行实例归一化和Relu激活。The feature transformation layer is followed by two convolution layers, and the convolution layers are used to restore the RGB image and realize the conversion of multiple features into RGB images; instance normalization and ReLU activation are performed after each convolution.
进一步,所述增强网络模块包括4种损失函数:Furthermore, the enhanced network module includes four loss functions:
1)内容损失:1) Content loss:
根据预先训练VGG-19网络的ReLU层产生的激活图定义内容损失;所述内容损失为增强图片和目标图片的特征表示之间的欧式距离:The content loss is defined based on the activation map produced by the ReLU layer of the pre-trained VGG-19 network; the content loss is the Euclidean distance between the feature representations of the enhanced image and the target image:
其中,φi为VGG-19网络在第i个卷积层之后获得的特征图;Among them, φ i is the feature map obtained by the VGG-19 network after the i-th convolutional layer;
2)总变差损失:2) Total variation loss:
其中,C,H,W分别是增强图片Ie的通道数,高度,宽度,分别是增强图片在x,y反向的梯度;Among them, C, H, and W are the number of channels, height, and width of the enhanced image I e , respectively. They are respectively to enhance the gradient of the image in the opposite direction of x and y;
3)颜色损失:3) Color loss:
Lcolor=||δ(G(Ix))-δ(Iy)||2; Lcolor =||δ(G( Ix ))-δ( Iy )|| 2 ;
其中,δ表示高斯模糊函数,用于移除图片的局部细节;Among them, δ represents the Gaussian blur function, which is used to remove local details of the image;
4)对抗损失4) Fighting Losses
其中,D代表判别网络,G代表生成网络,Ix,Iy分别表示低照度图片,自然照度图片。Among them, D represents the discriminant network, G represents the generative network, I x and I y represent low-light images and natural-light images, respectively.
进一步,所述添加有增强网络的生成器包括:Further, the generator with the enhanced network added includes:
所述生成器按顺序由1个卷积块、4个残差块以及2个卷积块组成;The generator consists of 1 convolution block, 4 residual blocks and 2 convolution blocks in sequence;
所述4个残差块用于保持高度/宽度恒定;The four residual blocks are used to keep the height/width constant;
所述生成器每次卷积之后进行实例正则化和ReLU激活;The generator performs instance regularization and ReLU activation after each convolution;
所述生成器最后一个卷积层是tanh激活,此外每个卷积层之后利用ReLU激活。The last convolutional layer of the generator is tanh activated, and ReLU activation is used after each convolutional layer.
进一步,所述判别器包含5个卷积层、1个全连接层和1个softmax层;Furthermore, the discriminator includes 5 convolutional layers, 1 fully connected layer and 1 softmax layer;
所述卷积层卷积核的尺寸从11缩小到3,特征通道数从3增大到192;用于逐步抽取输入特征;The size of the convolution kernel of the convolution layer is reduced from 11 to 3, and the number of feature channels is increased from 3 to 192; it is used to gradually extract input features;
所述全连接层和softmax层用于根据提取的特征图预测其来源于生成器或真实图片的可能,结果得到一个(Batch,Ptrue,Pfalse)3元组,Ptrue,Pfalse值都在[0,1]范围。The fully connected layer and the softmax layer are used to predict the possibility that the extracted feature map comes from the generator or the real picture, and the result is a (Batch, P true , P false ) 3-tuple, where the values of P true and P false are both in the range of [0, 1].
本发明另一目的在于提供一种计算机设备,所述计算机设备包括存储器和处理器,所述存储器存储有计算机程序,所述计算机程序被所述处理器执行时,使得所述处理器执行如下步骤:Another object of the present invention is to provide a computer device, the computer device comprising a memory and a processor, the memory storing a computer program, and when the computer program is executed by the processor, the processor performs the following steps:
摄影端利用摄影设备进行图像数据获取与对象捕捉,并将获取的图像数据传送至增强端;The photographing end uses photographic equipment to acquire image data and capture objects, and transmits the acquired image data to the enhancing end;
增强端利用构建的基于增强网络模块生成对抗网络的低照度图像增强模型对传输的低照度图像进行增强,并运用对抗网络算法进行图像识别,为未进行过识别的种类贴上标签;同时提取存储后台中已完成识别的类似标签进行辅助识别;The enhancement end uses the low-light image enhancement model built based on the enhancement network module to generate adversarial networks to enhance the transmitted low-light images, and uses the adversarial network algorithm to perform image recognition and label the types that have not been recognized. At the same time, similar labels that have been recognized in the storage background are extracted for auxiliary recognition.
将完成识别增强的图片传输至摄影端以及存储后台分别进行输出与存储。The images that have completed recognition and enhancement are transmitted to the camera end and the storage background for output and storage respectively.
本发明另一目的在于提供一种计算机可读存储介质,存储有计算机程序,所述计算机程序被处理器执行时,使得所述处理器执行如下步骤:Another object of the present invention is to provide a computer-readable storage medium storing a computer program, wherein when the computer program is executed by a processor, the processor executes the following steps:
摄影端利用摄影设备进行图像数据获取与对象捕捉,并将获取的图像数据传送至增强端;The photographing end uses photographic equipment to acquire image data and capture objects, and transmits the acquired image data to the enhancing end;
增强端利用构建的基于增强网络模块生成对抗网络的低照度图像增强模型对传输的低照度图像进行增强,并运用对抗网络算法进行图像识别,为未进行过识别的种类贴上标签;同时提取存储后台中已完成识别的类似标签进行辅助识别;The enhancement end uses the low-light image enhancement model built based on the enhancement network module to generate adversarial networks to enhance the transmitted low-light images, and uses the adversarial network algorithm to perform image recognition and label the types that have not been recognized. At the same time, similar labels that have been recognized in the storage background are extracted for auxiliary recognition.
将完成识别增强的图片传输至摄影端以及存储后台分别进行输出与存储The image after recognition enhancement is transmitted to the camera end and storage background for output and storage respectively
结合上述的所有技术方案,本发明所具备的优点及积极效果为:本发明基于增强网络模块生成对抗网络的低照度图像增强算法的刑侦摄影系统具有现阶段对于低照度图像曾强效果最好的增强网络模块生成对抗网络的低照度图像增强算法。本发明提供了一种成体系的刑侦摄影效果增强系统。本发明基于生成对抗网络的低照度图像增强算法于刑侦领域的应用Combining all the above technical solutions, the advantages and positive effects of the present invention are as follows: the criminal investigation photography system based on the low-light image enhancement algorithm of the enhancement network module generative adversarial network of the present invention has the best low-light image enhancement effect for the enhancement network module generative adversarial network at present. The present invention provides a systematic criminal investigation photography effect enhancement system. Application of the low-light image enhancement algorithm based on the generative adversarial network of the present invention in the field of criminal investigation
本发明低照度图像增强效果最好、增强效率高、应成体系的系统可应用于各种复杂场景、选取了最适用于刑侦摄影的算法技术。The low-illumination image enhancement effect of the present invention is the best, the enhancement efficiency is high, the system can be applied to various complex scenes, and the algorithm technology most suitable for criminal investigation photography is selected.
对比的技术效果或者实验效果包括:The technical effects or experimental effects of the comparison include:
低光配对数据集(LOL)包含500张低/正常光图像对,是一个在真实场景中用于低照度增强拍摄的图像对的数据集,大多数弱光图像是通过更改曝光时间和ISO来收集的,并使用了三步法对齐了图像对。该数据集包含从各种场景中捕获的图像,例如房屋,校园,俱乐部,街道。基于该数据集,本发明制作了11592张训练集照片,360张测试集照片。The Low Light Pairing Dataset (LOL) contains 500 low/normal light image pairs. It is a dataset of image pairs used for low-light enhanced shooting in real scenes. Most low-light images are collected by changing the exposure time and ISO, and the image pairs are aligned using a three-step method. The dataset contains images captured from various scenes, such as houses, campuses, clubs, and streets. Based on this dataset, the present invention produced 11,592 training set photos and 360 test set photos.
在LOL数据集上,本发明对比了HE,SRIE,DSLR这3种方法,由于有些方法无法实现去噪功能,本发明联合BM3D方法进行去噪,进而产生最终的结果。定量的结果如表3所示。On the LOL dataset, the present invention compares the three methods of HE, SRIE, and DSLR. Since some methods cannot achieve denoising function, the present invention combines BM3D method for denoising, and then produces the final result. The quantitative results are shown in Table 3.
表3本发明的方法同HE,SRIE,DSLR在LOL数据集实验结果Table 3 Experimental results of the method of the present invention with HE, SRIE, and DSLR on the LOL dataset
本发明的方法同HE,SRIE,DSLR在LOL数据集视觉效果中,定性的结果如图7(a)低照度图片一、7(b)现有技术HE、7(c)现有技术SRIE、7(d)现有技术DSLR、7(e)本发明、7(f)本发明处理后真实照片一。The method of the present invention is the same as HE, SRIE, and DSLR in the visual effect of the LOL dataset. The qualitative results are shown in Figure 7 (a) low-light picture 1, 7 (b) prior art HE, 7 (c) prior art SRIE, 7 (d) prior art DSLR, 7 (e) the present invention, and 7 (f) the real photo after processing by the present invention.
如图8(a)低照度图片二、图8(b)现有技术HE、图8(c)现有技术SRIE、图8(d)现有技术DSLR、图8(e)本发明、图8(f)本发明处理后真实照片二。As shown in Figure 8(a) low-light picture 2, Figure 8(b) prior art HE, Figure 8(c) prior art SRIE, Figure 8(d) prior art DSLR, Figure 8(e) the present invention, and Figure 8(f) real photo 2 after processing by the present invention.
对比图7、8中各算法的主观视觉效果可以看出,HE算法在LOL数据集上出现了较多的内容失真和颜色失真,例如增强第一张图片时,HE的图片中墙壁上了出现了明显背景无关的伪影,并且整体图像颜色偏灰,增强的第二张图片时,深色的地板和真实照片淡木色地板明显不符;SRIE算法增强后的图片则出现较多亮度失真,从第一张和第二张图片可以看出,其对低照度图片亮度的提升还不足以辨别图像内容;DSLR算法能改善图片亮度和保留一定内容,但仍存在过曝光和局部噪声问题,如增强的两张图片中,第一张白色的纳衣柜门明显偏亮,第二张图片的地板和墙壁亮度过高,同时放大图片后可以发现,其局部存在一定的噪声和细节不够平滑;而本发明的方法增强后的图片亮度适中,避免了过曝光问题,同时内容保留完整,细节信息丢失较少。By comparing the subjective visual effects of each algorithm in Figures 7 and 8, it can be seen that the HE algorithm has more content distortion and color distortion on the LOL dataset. For example, when enhancing the first picture, obvious background-irrelevant artifacts appear on the wall in the HE picture, and the overall image color is gray. When enhancing the second picture, the dark floor is obviously inconsistent with the light wooden floor in the real photo; the picture enhanced by the SRIE algorithm has more brightness distortion. It can be seen from the first and second pictures that its improvement in the brightness of low-light pictures is not enough to distinguish the image content; the DSLR algorithm can improve the brightness of the picture and retain certain content, but there are still problems of overexposure and local noise. For example, in the two enhanced pictures, the white wardrobe door in the first picture is obviously brighter, and the floor and wall brightness of the second picture are too high. At the same time, after enlarging the picture, it can be found that there is certain noise in the local area and the details are not smooth enough; while the picture enhanced by the method of the present invention has moderate brightness, avoiding the problem of overexposure, while the content is kept intact and less detail information is lost.
从表3可知,本发明的方法在2个指标上均超过了其他方法,表明在真实数据集上,本发明的方法性能也较优秀,鲁棒性较好。相比传统的方法HE,SRIE,本发明在PSNR上分别提高了36.11%,77.27%,在SSIM上分别提高了74.06%,86.76%;对比当前的深度学习方法DSLR,本发明在PSNR上也提升了9.69%,SSIM提升了0.04%。由此可见,本发明的方法较传统方法效果有很大提升,较深度学习的方法也有一定优势。As can be seen from Table 3, the method of the present invention outperforms other methods in both indicators, indicating that the method of the present invention also has excellent performance and good robustness on real data sets. Compared with the traditional methods HE and SRIE, the present invention improves PSNR by 36.11% and 77.27% respectively, and improves SSIM by 74.06% and 86.76% respectively; compared with the current deep learning method DSLR, the present invention also improves PSNR by 9.69% and SSIM by 0.04%. It can be seen that the method of the present invention has a greatly improved effect compared with the traditional method, and also has certain advantages over the deep learning method.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
为了更清楚地说明本申请实施例的技术方案,下面将对本申请实施例中所需要使用的附图做简单的介绍,显而易见地,下面所描述的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions of the embodiments of the present application, the following is a brief introduction to the drawings required for use in the embodiments of the present application. Obviously, the drawings described below are only some embodiments of the present application. For ordinary technicians in this field, other drawings can be obtained based on these drawings without creative work.
图1是本发明实施例提供的基于低照度图像增强算法的刑侦摄影系统结构示意图。FIG1 is a schematic diagram of the structure of a criminal investigation photography system based on a low-light image enhancement algorithm provided by an embodiment of the present invention.
图2是本发明实施例提供的基于低照度图像增强算法的刑侦摄影方法流程图。FIG. 2 is a flow chart of a criminal investigation photography method based on a low-light image enhancement algorithm provided in an embodiment of the present invention.
图3是本发明实施例提供的生成器网络结构示意图。FIG3 is a schematic diagram of a generator network structure provided in an embodiment of the present invention.
图4是本发明实施例提供的判别器网络结构示意图。FIG. 4 is a schematic diagram of a discriminator network structure provided by an embodiment of the present invention.
图5是本发明实施例提供的残差模块网络结构图。FIG5 is a diagram of a residual module network structure according to an embodiment of the present invention.
图6是本发明实施例提供的增强分支网络结构图。FIG. 6 is a diagram of an enhanced branch network structure provided by an embodiment of the present invention.
图7(a)本发明实施例提供的低照度图片一、7(b)现有技术HE、7(c)现有技术SRIE、7(d)现有技术DSLR、7(e)本发明、7(f)本发明处理后真实照片一。Figure 7(a) is a low-light picture 1 provided by an embodiment of the present invention, 7(b) is a prior art HE, 7(c) is a prior art SRIE, 7(d) is a prior art DSLR, 7(e) is the present invention, and 7(f) is a real photo 1 after being processed by the present invention.
图8(a)本发明实施例提供的低照度图片二、图8(b)现有技术HE、图8(c)现有技术SRIE、图8(d)现有技术DSLR、图8(e)本发明、图8(f)本发明处理后真实照片二。FIG8(a) is a second low-light picture provided by an embodiment of the present invention, FIG8(b) is HE in the prior art, FIG8(c) is SRIE in the prior art, FIG8(d) is DSLR in the prior art, FIG8(e) is the present invention, and FIG8(f) is a second real photo processed by the present invention.
具体实施方式Detailed ways
为了使本发明的目的、技术方案及优点更加清楚明白,以下结合实施例,对本发明进行进一步详细说明。应当理解,此处所描述的具体实施例仅仅用以解释本发明,并不用于限定本发明。In order to make the purpose, technical solution and advantages of the present invention more clearly understood, the present invention is further described in detail below in conjunction with the embodiments. It should be understood that the specific embodiments described herein are only used to explain the present invention and are not used to limit the present invention.
针对现有技术存在的问题,本发明提供了一种基于低照度图像增强算法的刑侦摄影系统,下面结合附图对本发明作详细的描述。In view of the problems existing in the prior art, the present invention provides a criminal investigation photography system based on a low-illumination image enhancement algorithm. The present invention is described in detail below in conjunction with the accompanying drawings.
如图1所示,本发明实施例提供的基于低照度图像增强算法的刑侦摄影系统包括:摄影端,增强端与储存后台;As shown in FIG1 , the criminal investigation photography system based on the low-light image enhancement algorithm provided by the embodiment of the present invention includes: a photography end, an enhancement end and a storage background;
所述摄影端,与增强端连接;包括摄影模块与输出模块;所述摄影模块,用于进行图像数据获取与对象捕捉,并将获取的图像数据传送至增强端;所述输出模块用于输出增强后的图像;The photographing end is connected to the enhancing end; it includes a photographing module and an output module; the photographing module is used to acquire image data and capture objects, and transmit the acquired image data to the enhancing end; the output module is used to output the enhanced image;
所述增强端,与摄影端以及储存后台连接,包括增强模块,识别模块与传输模块;The enhancement end is connected to the photographing end and the storage background, and includes an enhancement module, a recognition module and a transmission module;
所述增强模块,用于利用基于增强网络模块生成对抗网络的低照度图像增强模型对传输的低照度图像进行增强;所述识别模块,用于运用对抗网络算法进行图像识别;同时用于提取后台中已完成识别的类似标签进行辅助识别,也可给未进行过识别的种类贴上标签;所述传输模块,用于将完成识别增强的图片传输至摄影端中的输出模块,同时用于将增强、识别结果与储存后台中未进行过存储的标签传输至后台进行保存;The enhancement module is used to enhance the transmitted low-light image by using a low-light image enhancement model based on the enhancement network module to generate an adversarial network; the recognition module is used to perform image recognition using an adversarial network algorithm; it is also used to extract similar tags that have been recognized in the background for auxiliary recognition, and it can also label types that have not been recognized; the transmission module is used to transmit the image that has completed recognition and enhancement to the output module in the camera end, and is also used to transmit the enhancement and recognition results and tags that have not been stored in the storage background to the background for storage;
所述存储后台,与增强端连接,用于保存增强、识别结果与识别种类标签。The storage backend is connected to the enhancement end and is used to store enhancement, recognition results and recognition category labels.
如图2所示,本发明实施例提供的基于低照度图像增强算法的刑侦摄影方法包括:As shown in FIG2 , the criminal investigation photography method based on the low-light image enhancement algorithm provided by the embodiment of the present invention includes:
S101,摄影端利用摄影设备进行图像数据获取与对象捕捉,并将获取的图像数据传送至增强端;S101, the photographing end uses a photographing device to acquire image data and capture an object, and transmits the acquired image data to the enhancing end;
S102,增强端利用构建的基于增强网络模块生成对抗网络的低照度图像增强模型对传输的低照度图像进行增强,并运用对抗网络算法进行图像识别,为未进行过识别的种类贴上标签;同时提取存储后台中已完成识别的类似标签进行辅助识别;S102, the enhancement end uses the low-light image enhancement model based on the enhancement network module to generate an adversarial network to enhance the transmitted low-light image, and uses the adversarial network algorithm to perform image recognition, and labels the types that have not been recognized; at the same time, similar labels that have been recognized in the storage background are extracted for auxiliary recognition;
S103,将完成识别增强的图片传输至摄影端以及存储后台分别进行输出与存储。S103, transmitting the image after recognition enhancement to the camera end and the storage background for output and storage respectively.
步骤S101中,会根据摄影端硬件设备的不同,如不同相机的不同的特点,以及所需求捕捉的对象不同,进行硬件设备功能的特殊设置。其捕捉与传输技术已广泛的应用于各种类型的智能设备。In step S101, special settings of the hardware device functions are performed according to the different hardware devices at the photography end, such as the different characteristics of different cameras, and the different objects to be captured. Its capture and transmission technology has been widely used in various types of smart devices.
其捕捉技术大体流程如下:The capture technology generally follows the following process:
生成camera controller(各类型的摄影控制系统)作为操作引擎接口。Generate camera controller (various types of photography control systems) as the operation engine interface.
初始化摄影设备引擎。Initialize the camera device engine.
进行异步操开始掌控相机整体。Perform asynchronous operations to start controlling the entire camera.
异步操作启动相机,同时调用相机观察系统。The asynchronous operation starts the camera and calls the camera observation system.
利用相机观察系统开始准备图像捕捉,同时设置引擎为空闲状态。Start preparing for image capture using the camera observation system and set the engine to idle state.
需要拍摄图片后,调用相机控制系统,执行引擎中的工程命令,之后进行操作。When you need to take a picture, call the camera control system, execute the engineering commands in the engine, and then perform the operation.
进行捕捉图片异步操作,完成后调用相机引擎中的图片捕捉完成指令完成操作。Perform asynchronous image capture operations. After completion, call the image capture completion instruction in the camera engine to complete the operation.
步骤S103中,图像传输已经广泛的运用于各种类型的智能设备。在图像进行传输,并且完成储存这个过程中,本发明需要完成图像数字化处理,凭借此来存储图像的数字数据。In step S103, image transmission has been widely used in various types of intelligent devices. In the process of image transmission and storage, the present invention needs to complete image digitization processing to store the digital data of the image.
其数字化处理大体流程如下:The general process of digital processing is as follows:
采样采样的实质就是要用多少点来描述一幅图像,采样结果质量的高低就是用前面所说的图像分辨率来衡量.The essence of sampling is how many points are used to describe an image, and the quality of the sampling result is measured by the image resolution mentioned above.
量化量化是指要使用多大范围的数值来表示图像采样之后的每一个点.量化的结果是图像能够容纳的颜色总数,它反映了采样的质量.Quantization Quantization refers to the range of values used to represent each point after the image is sampled. The result of quantization is the total number of colors that the image can accommodate, which reflects the quality of sampling.
压缩编码数字化后得到的图像数据量十分巨大,必须采用编码技术来压缩其信息量。在一定意义上讲,编码压缩技术是实现图像传输与储存的关键。The amount of image data obtained after compression coding is very huge, and coding technology must be used to compress its information. In a certain sense, coding compression technology is the key to realizing image transmission and storage.
步骤S102中,本发明实施例提供的基于增强网络模块生成对抗网络的低照度图像增强模型构建方法包括:In step S102, the method for constructing a low-light image enhancement model based on an enhancement network module generating an adversarial network provided by an embodiment of the present invention includes:
(1)获取高质量的图像对作为训练数据集训练生成器网络G;(1) Obtain high-quality image pairs as training datasets to train the generator network G;
(2)从训练数据集随机采样m个低照度图片对其中,m表示训练批次的大小,Ix表示低照度图片,Iy表示真实照度图片,Iadv表示判别器的输入;(2) Randomly sample m low-light image pairs from the training dataset Where m represents the size of the training batch, I x represents the low-light image, I y represents the real-light image, and I adv represents the input of the discriminator;
(3)固定判别网络的输入为Iadv={0,0,…,0},长度为m;(3) The input of the discriminant network is fixed to I adv = {0,0,…,0}, with a length of m;
(4)最小化生成器网络总体损失:(4) Minimize the overall loss of the generator network:
Lossgen=ωaLa+ωadvLadv+ωconLcon+ωtvLtv+ωcolLcol;Loss gen =ω a L a +ω adv L adv +ω con L con +ω tv L tv +ω col L col ;
(5)训练判别器网络,随机初始化判别网络的输入为Iadv={1,0,…,0},长度为m;(5) Train the discriminator network and randomly initialize the input of the discriminator network to I adv = {1, 0, …, 0}, with a length of m;
(6)从训练数据集随机采样m个低照度图片 (6) Randomly sample m low-light images from the training dataset
(7)最大化判别器网络总体损失: (7) Maximize the overall loss of the discriminator network:
本发明实施例提供的基于增强网络模块生成对抗网络的低照度图像增强模型包括:The low-light image enhancement model based on the enhancement network module generative adversarial network provided by the embodiment of the present invention includes:
所述低照度图像增强模型包括添加有增强网络的生成器、判别器以及损失函数;The low-light image enhancement model includes a generator, a discriminator and a loss function added with an enhancement network;
所述添加有增强网络的生成器,基于完全卷积网络,由多个残差块和卷积块2部分组成;用于将输入图像作为一个整体转换成一个在新空间中类似的图片;The generator with the enhanced network added is based on a fully convolutional network and consists of two parts: a plurality of residual blocks and a convolutional block; it is used to convert the input image as a whole into a similar picture in a new space;
所述判别器,用于同时接收生成器生成的图片和真实图片,产生真假的预测值;The discriminator is used to simultaneously receive the image generated by the generator and the real image, and generate a true or false prediction value;
所述损失函数为:Loss=ωaLa+ωadvLadv+ωconLcon+ωtvLtv+ωcolLcol;The loss function is: Loss = ω a L a + ω adv L adv + ω con L con + ω tv L tv + ω col L col ;
其中La,Ladv,Lcon,Ltv,Lcol分别表示注意力损失,对抗损失,内容损失,总变差损失,颜色损失,ωa,ωadv,ωcon,ωtv,ωcol分别表示其损失对应权重。Among them, La , Ladv , Lcon , Ltv , Lcol represent attention loss, adversarial loss, content loss, total variation loss, and color loss respectively, and ωa , ωadv , ωcon , ωtv , ωcol represent the corresponding weights of their losses respectively.
本发明实施例提供的增强网络模块包括:The enhanced network module provided in the embodiment of the present invention includes:
所述增强网络模块,所述增强网络模块包含2层3x3卷积层以及特征变换层;用于克服图片低对比度的缺点并改善细节,进行图片效果增强;The enhanced network module includes two 3x3 convolutional layers and a feature transformation layer; it is used to overcome the disadvantage of low contrast of the image and improve the details to enhance the image effect;
所述3x3卷积层第一层用于进行特征提取,实现从rgb通道到多个特征;The first layer of the 3x3 convolutional layer is used for feature extraction, realizing from RGB channels to multiple features;
所述卷积层第一层之后为特征变换层,所述特征变换层为残差模块;所述特征变换层,用于通过连接多个残差单元进行复杂的特征变换;The first convolutional layer is followed by a feature transformation layer, which is a residual module; the feature transformation layer is used to perform complex feature transformation by connecting multiple residual units;
所述特征变换层后为两层卷积层,所述卷积层用于恢复rgb图片,实现多特征转换为rgb图片;所述每次卷积之后进行实例归一化和Relu激活。The feature transformation layer is followed by two convolution layers, and the convolution layers are used to restore the RGB image and realize the conversion of multiple features into RGB images; instance normalization and ReLU activation are performed after each convolution.
本发明实施例提供的增强网络模块包括4种损失函数:The enhanced network module provided in the embodiment of the present invention includes four loss functions:
1)内容损失:1) Content loss:
根据预先训练VGG-19网络的ReLU层产生的激活图定义内容损失;所述内容损失为增强图片和目标图片的特征表示之间的欧式距离:The content loss is defined based on the activation map produced by the ReLU layer of the pre-trained VGG-19 network; the content loss is the Euclidean distance between the feature representations of the enhanced image and the target image:
其中,φi为VGG-19网络在第i个卷积层之后获得的特征图;Among them, φ i is the feature map obtained by the VGG-19 network after the i-th convolutional layer;
2)总变差损失:2) Total variation loss:
其中,C,H,W分别是增强图片Ie的通道数,高度,宽度,分别是增强图片在x,y反向的梯度;Among them, C, H, and W are the number of channels, height, and width of the enhanced image I e , respectively. They are respectively to enhance the gradient of the image in the opposite direction of x and y;
3)颜色损失:3) Color loss:
Lcolor=||δ(G(Ix))-δ(Iy)||2; Lcolor =||δ(G( Ix ))-δ( Iy )|| 2 ;
其中,δ表示高斯模糊函数,用于移除图片的局部细节;Among them, δ represents the Gaussian blur function, which is used to remove local details of the image;
4)对抗损失4) Fighting Losses
其中,D代表判别网络,G代表生成网络,Ix,Iy分别表示低照度图片,自然照度图片。Among them, D represents the discriminant network, G represents the generative network, I x and I y represent low-light images and natural-light images, respectively.
本发明实施例提供的添加有增强网络的生成器包括:The generator with an enhanced network provided in an embodiment of the present invention includes:
所述生成器按顺序由1个卷积块、4个残差块以及2个卷积块组成;The generator consists of 1 convolution block, 4 residual blocks and 2 convolution blocks in sequence;
所述4个残差块用于保持高度/宽度恒定;The four residual blocks are used to keep the height/width constant;
所述生成器每次卷积之后进行实例正则化和ReLU激活;The generator performs instance regularization and ReLU activation after each convolution;
所述生成器最后一个卷积层是tanh激活,此外每个卷积层之后利用ReLU激活。The last convolutional layer of the generator is tanh activated, and ReLU activation is used after each convolutional layer.
本发明实施例提供的判别器包含5个卷积层、1个全连接层和1个softmax层;The discriminator provided in the embodiment of the present invention includes 5 convolutional layers, 1 fully connected layer and 1 softmax layer;
所述卷积层卷积核的尺寸从11缩小到3,特征通道数从3增大到192;用于逐步抽取输入特征;The size of the convolution kernel of the convolution layer is reduced from 11 to 3, and the number of feature channels is increased from 3 to 192; it is used to gradually extract input features;
所述全连接层和softmax层用于根据提取的特征图预测其来源于生成器或真实图片的可能,结果得到一个(Batch,Ptrue,Pfalse)3元组,Ptrue,Pfalse值都在[0,1]范围。The fully connected layer and the softmax layer are used to predict the possibility that the extracted feature map comes from the generator or the real picture, and the result is a (Batch, P true , P false ) 3-tuple, where the values of P true and P false are both in the range of [0, 1].
下面结合具体实施例对本发明的技术效果作进一步描述。The technical effects of the present invention are further described below in conjunction with specific embodiments.
实施例1:Embodiment 1:
4.1基于增强网络模块生成对抗网络的低照度图像增强算法框架4.1 Low-light image enhancement algorithm framework based on enhanced network module generative adversarial network
(1)网络结构(1) Network structure
低照度图像增强的网络结构如图3,图4所示,为了增强低照度图像中弱光照位置信息在网络流中的传递,生成器添加增强网络。增强网络将输入图像作为一个整体转换成一个在新空间中类似的图片,注意力网络预测弱光照的位置掩码,它和输入图像尺寸一样,每个像素点是0到1之间的概率值。最后本发明组合输入图像,注意力图和转换后的输入构成最终增强后的图像。判别器可以同时接收生成器生成的图片和真实图片,最后产生真假的预测值。The network structure of low-light image enhancement is shown in Figures 3 and 4. In order to enhance the transmission of low-light position information in the network flow of low-light images, the generator adds an enhancement network. The enhancement network converts the input image as a whole into a similar picture in the new space. The attention network predicts the position mask of the low-light, which is the same size as the input image, and each pixel is a probability value between 0 and 1. Finally, the present invention combines the input image, the attention map and the converted input to form the final enhanced image. The discriminator can receive the picture generated by the generator and the real picture at the same time, and finally generate a true or false prediction value.
(i)增强网络的网络结构细节如表1所示,基于完全卷积网络(FCN),并利用卷积神经网络的属性,例如翻译不变性和参数共享。网络由多个残差块和卷积块2部分组成。最开始是1个卷积块。中间部分包含4个残差块,保持高度/宽度恒定,每次卷积之后进行实例正则化和ReLU激活。最后是2个卷积块。除残差块外,最后一个卷积层是tanh激活,此外每个卷积层之后只有ReLU激活。(i) The network structure details of the enhanced network are shown in Table 1. It is based on a fully convolutional network (FCN) and exploits the properties of convolutional neural networks, such as translation invariance and parameter sharing. The network consists of two parts: multiple residual blocks and convolutional blocks. There is one convolutional block at the beginning. The middle part contains 4 residual blocks, keeping the height/width constant, and performing instance normalization and ReLU activation after each convolution. Finally, there are 2 convolutional blocks. In addition to the residual block, the last convolution layer is tanh activated, and there are only ReLU activations after each convolution layer.
表1增强网络的网络结构细节Table 1 Network structure details of the enhanced network
(ii)判别器网络的网络结构细节如表2所示,包含5个卷积层,1个全连接层和1个softmax层。多个卷积层用于逐步抽取输入特征,卷积核的尺寸从11缩小到3,特征通道数从3增大到192,对于低照度图片而言,由于光照分布不均以及噪声等影响,图像呈现出大面积暗黑,弱光等,导致局部特征单一,一开始大感受野有利于局部特征图获取更多信息,随着通道数的增加,特征逐渐丰富,此时小感受野有利于提取图片更多细节特征。全连接层和softmax层用于根据提取的特征图预测其来源于生成器或真实图片的可能,结果是一个(Batch,Ptrue,Pfalse)3元组,Ptrue,Pfalse值都在[0,1]范围。(ii) The details of the network structure of the discriminator network are shown in Table 2, which includes 5 convolutional layers, 1 fully connected layer and 1 softmax layer. Multiple convolutional layers are used to gradually extract input features. The size of the convolution kernel is reduced from 11 to 3, and the number of feature channels is increased from 3 to 192. For low-light images, due to the uneven distribution of light and the influence of noise, the image presents a large area of darkness and low light, resulting in a single local feature. At the beginning, a large receptive field is conducive to obtaining more information from the local feature map. As the number of channels increases, the features are gradually enriched. At this time, a small receptive field is conducive to extracting more detailed features of the image. The fully connected layer and the softmax layer are used to predict the possibility of the extracted feature map coming from the generator or the real picture. The result is a (Batch, P true , P false ) 3-tuple, and the values of P true and P false are both in the range of [0,1].
表2判别器网络的网络结构细节Table 2 Network structure details of the discriminator network
(2)损失函数(2) Loss Function
由于输入和目标照片无法紧密匹配(即像素到像素),即不同的光学元件和传感器会导致特定的局部非线性失真和像差,即使精确对齐后,每个图像对之间的像素数也会存在非恒定偏移。因此,标准的每像素损失,除了感知质量指标,不适用于本发明的情况。为了定性和定量地提高图像整体质量,本发明通过进一步提出新的损失函数:Since the input and target photos cannot be closely matched (i.e., pixel-to-pixel), i.e., different optical elements and sensors will cause specific local nonlinear distortions and aberrations, there will be a non-constant offset in the number of pixels between each image pair even after precise alignment. Therefore, standard per-pixel losses, except for perceptual quality indicators, are not applicable to the case of the present invention. In order to improve the overall image quality qualitatively and quantitatively, the present invention further proposes a new loss function:
Loss=ωaLa+ωadvLadv+ωconLcon+ωtvLtv+ωcolLcol (4-1)Loss= ωaLa + ωadvLadv + ωconLcon + ωtvLtv + ωcolLcol ( 4-1 )
其中La,Ladv,Lcon,Ltv,Lcol分别表示注意力损失,对抗损失,内容损失,总变差损失,颜色损失,ωa,ωadv,ωcon,ωtv,ωcol分别表示其损失对应权重。Among them, La , Ladv , Lcon , Ltv , Lcol represent attention loss, adversarial loss, content loss, total variation loss, and color loss respectively, and ωa , ωadv , ωcon , ωtv , ωcol represent the corresponding weights of their losses respectively.
(3)算法流程图(3) Algorithm flow chart
训练数据生成后,本发明使用了高质量的图像对来反复训练本发明的GAN网络。在训练判别器阶段,本发明将一个批次的生成样本和一个批次的真实样本随机混淆,生成一个批次的混淆样本重新作为判别器输入。判别器试图识别真实和伪造的图像,因此对判别器进行了训练,其等价于最大化判别损失的过程。训练生成器网络的过程是使公式3-1最小化,以保证生成图片相较真实图片各方面损失最小,生成效果逼真。After the training data is generated, the present invention uses high-quality image pairs to repeatedly train the GAN network of the present invention. In the stage of training the discriminator, the present invention randomly confuses a batch of generated samples with a batch of real samples, generates a batch of confused samples and re-inputs them as the discriminator. The discriminator attempts to identify real and fake images, so the discriminator is trained, which is equivalent to the process of maximizing the discriminant loss. The process of training the generator network is to minimize formula 3-1 to ensure that the generated image has the least loss in all aspects compared to the real image, and the generated effect is realistic.
为更简洁明了地表示整个算法流程,G,D分别表示为生成器网络,判别器网络,训练时一个批次的大小为m。具体细节参照以下算法流程:To express the entire algorithm flow more concisely and clearly, G and D represent the generator network and the discriminator network respectively, and the size of a batch during training is m. For specific details, refer to the following algorithm flow:
4.2增强网络模块4.2 Enhanced Network Module
该模块作为主干网络,用于克服图片低对比度的缺点并改善细节,实现增强图片的效果。考虑到低照度图片中弱光区域特征信息较少,且可能存在噪声干扰,本发明使用残差连接构建基础残差模块,用于加深网络层数,提高网络对低照度图片增强的建模能力,避免由于网络加深造成特征损失。因此,本发明采用了残差模块作为该增强网络中的特征变换层,如图5所示。This module is used as a backbone network to overcome the disadvantages of low contrast in images and improve details to achieve the effect of enhancing images. Considering that there is less feature information in low-light areas of low-light images and there may be noise interference, the present invention uses residual connections to construct basic residual modules to deepen the number of network layers, improve the network's modeling capabilities for low-light image enhancement, and avoid feature loss due to network deepening. Therefore, the present invention uses a residual module as the feature transformation layer in the enhancement network, as shown in Figure 5.
如图5所示,该模块包含2层3x3普通卷积,每次卷积之后接着进行实例归一化和Relu激活,最后和输入相加得到最终输出结果。残差网络己经在多个领域证实了其在像素级任务中的有效性,如目标检测、语义及图像分割。该操作使用残差连接,可以训练更深的网络,并避免特征损失。As shown in Figure 5, the module contains 2 layers of 3x3 ordinary convolutions, followed by instance normalization and ReLU activation after each convolution, and finally added to the input to get the final output result. Residual networks have proven their effectiveness in pixel-level tasks in many fields, such as object detection, semantics, and image segmentation. This operation uses residual connections, which can train deeper networks and avoid feature loss.
通过引入残差摸块,增强网络结构如图6所示。By introducing the residual module, the enhanced network structure is shown in Figure 6.
如图6所示,第一层用于特征提取,实现从rgb通道到多个特征,最后两层用于恢复rgb图片,实现多特征转换为rgb图片;图中橙色块表示残差单元,通过连接多个残差单元进行复杂的特征变换,提高网络增强低照度图片的建模能力。As shown in Figure 6, the first layer is used for feature extraction to realize the conversion from RGB channels to multiple features. The last two layers are used to restore RGB images and realize the conversion of multiple features into RGB images. The orange blocks in the figure represent residual units. By connecting multiple residual units to perform complex feature transformations, the network's modeling ability for enhanced low-light images is improved.
为了约束增强网络提高图片整体感知质量,设计了如下4种损失函数:In order to constrain the enhancement network to improve the overall perceived quality of the image, the following four loss functions are designed:
(i)内容损失(i) Content loss
本发明根据预先训练VGG-19网络的ReLU层产生的激活图来定义内容损失。这种损失鼓励它们具有相似的特征表示,而不是测量图像之间的每像素差异,而是包含内容和感知质量的各个方面。在本发明的情况下,它用于保存图片语义,因为其他损失都没有考虑。令φi为VGG-19网络在第i个卷积层之后获得的特征图,则本发明的内容损失定义为增强图片和目标图片的特征表示之间的欧式距离The present invention defines a content loss based on the activation maps produced by the ReLU layers of a pre-trained VGG-19 network. This loss encourages them to have similar feature representations, and instead of measuring per-pixel differences between images, it encompasses aspects of content and perceptual quality. In the case of the present invention, it is used to preserve image semantics, as none of the other losses take this into account. Let φ i be the feature map obtained by the VGG-19 network after the i-th convolutional layer, then the content loss of the present invention is defined as the Euclidean distance between the feature representations of the enhanced image and the target image
(ii)总变差损失(ii) Total variation loss
这个损失可以增强图像的空间平滑度,对生成的组合图像的像素进行操作,促使生成图像具有空间连续性,从而避免相邻像素之间的重大差异,防止图像中出现棋盘图案。This loss can enhance the spatial smoothness of the image and operate on the pixels of the generated combined image to force the generated image to have spatial continuity, thereby avoiding significant differences between adjacent pixels and preventing the appearance of checkerboard patterns in the image.
这里的C,H,W分别是增强图片Ie的通道数,高度,宽度,分别是增强图片在x,y反向的梯度。由于他的权重相对较低,在消除噪声的同时也能不损害图片高频部分。Here C, H, and W are the number of channels, height, and width of the enhanced image I e , respectively. They enhance the image's gradient in the opposite direction of x and y. Since their weights are relatively low, they can remove noise without damaging the high-frequency part of the image.
(iii)颜色损失(iii) Color loss
为了避免颜色失真,同时评估增强后图片和目标图片间的色彩差异,本发明引入颜色损失。它是先对图片应用高斯模糊,再计算他们的欧式距离。颜色损失定义如下:In order to avoid color distortion and evaluate the color difference between the enhanced image and the target image, the present invention introduces color loss. It first applies Gaussian blur to the image and then calculates their Euclidean distance. The color loss is defined as follows:
Lcolor=||δ(G(Ix))-δ(Iy)||2 (3-4)L color =||δ(G(I x ))-δ(I y )|| 2 (3-4)
这里δ表示高斯模糊函数,用于移除图片的局部细节,但保留其全局信息,例如色彩。Here δ represents a Gaussian blur function, which is used to remove local details of the image but retain its global information, such as color.
(v)对抗损失(v) Fighting Losses
该损失鼓励生成网络转换低照度图片到自然图片,促进生成器学习自然图像的特性,包括纹理,对比度等。同时本发明使用梯度惩罚来稳定判别器的训练的,对抗损失定义如下:This loss encourages the generative network to convert low-light images into natural images, and promotes the generator to learn the characteristics of natural images, including texture, contrast, etc. At the same time, the present invention uses gradient penalty to stabilize the training of the discriminator, and the adversarial loss is defined as follows:
这里D代表判别网络,G代表生成网络,Ix,Iy分别表示低照度图片,自然照度图片。Here D represents the discriminant network, G represents the generative network, I x and I y represent low-light images and natural-light images, respectively.
4.3基于增强网络模块生成对抗网络的低照度图像增强算法的刑侦摄影系统模型4.3 Criminal Investigation Photography System Model Based on Low-Illumination Image Enhancement Algorithm Generative Adversarial Network Enhancement Network Module
本发明将整个刑侦摄影系统分为三个主要部分,摄影端,增强端与储存后台。The present invention divides the entire criminal investigation photography system into three main parts: a photography end, an enhancement end and a storage background.
在摄影端,分为两部分,分别为摄影模块与输出模块。摄影模块进行照片拍摄与对象捕捉的功能,与增强段相连,在完成拍摄之后将照片传到增强端进行进一步处理。输出模块与增强端相连,负责输出增强后的图像。On the photography side, it is divided into two parts, namely the photography module and the output module. The photography module performs the functions of photo shooting and object capture, and is connected to the enhancement segment. After shooting, the photo is transmitted to the enhancement side for further processing. The output module is connected to the enhancement side and is responsible for outputting the enhanced image.
增强端由增强模块,识别模块与传输模块组成。增强模块内含本发明上述内容中的增强算法,负责对传输过来的低照度图像进行增强。识别模块则运用对抗网络算法进行图像识别,并且与储存后台相连,可以提取后台中已完成识别的类似标签进行辅助识别,也可以给未进行过识别的种类贴上标签。传输模块则负责将完成识别增强的图片传输至摄影端中的输出模块,也负责将增强,识别结果与储存后台中未进行过存储的标签传输至后台进行保存。The enhancement end is composed of an enhancement module, a recognition module and a transmission module. The enhancement module contains the enhancement algorithm described above in the present invention, and is responsible for enhancing the low-light image transmitted. The recognition module uses an adversarial network algorithm for image recognition, and is connected to the storage background. It can extract similar tags that have been identified in the background for auxiliary recognition, and can also label types that have not been identified. The transmission module is responsible for transmitting the image that has completed the recognition enhancement to the output module in the photography end, and is also responsible for transmitting the enhancement, recognition results and tags that have not been stored in the storage background to the background for storage.
存储后台,负责保存增强,识别结果与识别种类标签。The storage backend is responsible for saving enhancement, recognition results and recognition category labels.
下面结合具体实验效果对本发明作进一步描述。The present invention is further described below in conjunction with specific experimental results.
本发明的方法同HE,SRIE,DSLR在LOL数据集视觉效果中,定性的结果如图7(a)低照度图片一、7(b)现有技术HE、7(c)现有技术SRIE、7(d)现有技术DSLR、7(e)本发明、7(f)本发明处理后真实照片一。The method of the present invention is the same as HE, SRIE, and DSLR in the visual effect of the LOL dataset. The qualitative results are shown in Figure 7 (a) low-light picture 1, 7 (b) prior art HE, 7 (c) prior art SRIE, 7 (d) prior art DSLR, 7 (e) the present invention, and 7 (f) the real photo after processing by the present invention.
如图8(a)低照度图片二、图8(b)现有技术HE、图8(c)现有技术SRIE、图8(d)现有技术DSLR、图8(e)本发明、图8(f)本发明处理后真实照片二。As shown in Figure 8(a) low-light picture 2, Figure 8(b) prior art HE, Figure 8(c) prior art SRIE, Figure 8(d) prior art DSLR, Figure 8(e) the present invention, and Figure 8(f) real photo 2 after processing by the present invention.
以上所述,仅为本发明的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,凡在本发明的精神和原则之内所作的任何修改、等同替换和改进等,都应涵盖在本发明的保护范围之内。The above description is only a specific implementation mode of the present invention, but the protection scope of the present invention is not limited thereto. Any modifications, equivalent substitutions and improvements made by any technician familiar with the technical field within the technical scope disclosed by the present invention and within the spirit and principle of the present invention should be covered by the protection scope of the present invention.
Claims (7)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202010753913.6A CN111899193B (en) | 2020-07-30 | 2020-07-30 | Criminal investigation photographing system and method based on low-illumination image enhancement algorithm |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202010753913.6A CN111899193B (en) | 2020-07-30 | 2020-07-30 | Criminal investigation photographing system and method based on low-illumination image enhancement algorithm |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN111899193A CN111899193A (en) | 2020-11-06 |
| CN111899193B true CN111899193B (en) | 2024-06-14 |
Family
ID=73184081
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202010753913.6A Active CN111899193B (en) | 2020-07-30 | 2020-07-30 | Criminal investigation photographing system and method based on low-illumination image enhancement algorithm |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN111899193B (en) |
Families Citing this family (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN112381897B (en) * | 2020-11-16 | 2023-04-07 | 西安电子科技大学 | Low-illumination image enhancement method based on self-coding network structure |
| CN112669284A (en) * | 2020-12-29 | 2021-04-16 | 天津大学 | Method for realizing pulmonary nodule detection by generating confrontation network |
| CN115187464A (en) * | 2021-04-02 | 2022-10-14 | 深圳市奥拓电子股份有限公司 | Method, device and system for repairing and denoising Micro LED display image |
| CN113436118B (en) * | 2021-08-10 | 2022-09-27 | 安徽工程大学 | Low-dose CT image restoration method based on multi-scale convolutional coding network |
Citations (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN108710831A (en) * | 2018-04-24 | 2018-10-26 | 华南理工大学 | A kind of small data set face recognition algorithms based on machine vision |
Family Cites Families (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106875352B (en) * | 2017-01-17 | 2019-08-30 | 北京大学深圳研究生院 | A low-light image enhancement method |
| KR102641116B1 (en) * | 2018-08-23 | 2024-02-27 | 삼성전자주식회사 | Method and device to recognize image and method and device to train recognition model based on data augmentation |
| CN109636754B (en) * | 2018-12-11 | 2022-05-31 | 山西大学 | Generative Adversarial Network-Based Very Low Illumination Image Enhancement Method |
| KR102134405B1 (en) * | 2019-06-27 | 2020-07-15 | 중앙대학교 산학협력단 | System and Method for Improving Low Light Level Image Using Generative Adversarial Network |
| CN110889813B (en) * | 2019-11-15 | 2023-11-24 | 安徽大学 | Low-light image enhancement method based on infrared information |
| CN111161178A (en) * | 2019-12-25 | 2020-05-15 | 湖南大学 | Single low-light image enhancement method based on generation type countermeasure network |
| CN111242865A (en) * | 2020-01-10 | 2020-06-05 | 南京航空航天大学 | Fundus image enhancement method based on generation type countermeasure network |
-
2020
- 2020-07-30 CN CN202010753913.6A patent/CN111899193B/en active Active
Patent Citations (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN108710831A (en) * | 2018-04-24 | 2018-10-26 | 华南理工大学 | A kind of small data set face recognition algorithms based on machine vision |
Non-Patent Citations (1)
| Title |
|---|
| 基于深度图像增强的夜间车道线检测技术;宋扬;李竹;;计算机应用;20191230(第S2期);全文 * |
Also Published As
| Publication number | Publication date |
|---|---|
| CN111899193A (en) | 2020-11-06 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN111899193B (en) | Criminal investigation photographing system and method based on low-illumination image enhancement algorithm | |
| CN113658057B (en) | Swin converter low-light-level image enhancement method | |
| CN111915526B (en) | Photographing method of low-illumination image enhancement algorithm based on brightness attention mechanism | |
| CN111784578B (en) | Image processing, model training method and device, equipment, storage medium | |
| CN109671023B (en) | A Super-resolution Reconstruction Method of Face Image | |
| Wang et al. | MAGAN: Unsupervised low-light image enhancement guided by mixed-attention | |
| CN111292264A (en) | A Deep Learning-Based Image High Dynamic Range Reconstruction Method | |
| CN111915525A (en) | Low-illumination image enhancement method based on improved depth separable generation countermeasure network | |
| CN117058019B (en) | A low-light target detection method based on pyramid enhancement network | |
| CN111489321B (en) | Deep Network Image Enhancement Method and System Based on Derivative Graph and Retinex | |
| CN109389569B (en) | Monitoring video real-time defogging method based on improved DehazeNet | |
| CN115661012B (en) | A multi-exposure image fusion system based on global-local aggregation learning | |
| CN114638764B (en) | Multi-exposure image fusion method and system based on artificial intelligence | |
| CN112508812A (en) | Image color cast correction method, model training method, device and equipment | |
| CN111242868A (en) | Image enhancement method based on convolutional neural network in dark vision environment | |
| CN115311149A (en) | Image denoising method, model, computer-readable storage medium and terminal device | |
| CN110827375A (en) | A method and system for true color coloring of infrared images based on low-light images | |
| CN113643202B (en) | Low-light-level image enhancement method based on noise attention-seeking instruction | |
| CN115457265A (en) | Image defogging method and system based on generation countermeasure network and multi-scale fusion | |
| CN120259101B (en) | Method and system for fusing inspection infrared image and visible light image of power transmission and transformation equipment | |
| CN117115018A (en) | A low-light image enhancement method and device considering real noise and artifacts | |
| Barua et al. | Arthdr-net: Perceptually realistic and accurate hdr content creation | |
| WO2025194622A1 (en) | Low-illuminance image enhancement method based on virtual exposure and dual fusion | |
| WO2022257184A1 (en) | Method for acquiring image generation apparatus, and image generation apparatus | |
| Tan et al. | Two‐Stage CNN Model for Joint Demosaicing and Denoising of Burst Bayer Images |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant | ||
| OL01 | Intention to license declared | ||
| OL01 | Intention to license declared |