CN105005788A - Target perception method based on emulation of human low level vision - Google Patents
Target perception method based on emulation of human low level vision Download PDFInfo
- Publication number
- CN105005788A CN105005788A CN201510377158.5A CN201510377158A CN105005788A CN 105005788 A CN105005788 A CN 105005788A CN 201510377158 A CN201510377158 A CN 201510377158A CN 105005788 A CN105005788 A CN 105005788A
- Authority
- CN
- China
- Prior art keywords
- fixation
- target
- area
- pixel
- saliency
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/19—Recognition using electronic means
- G06V30/192—Recognition using electronic means using simultaneous comparisons or correlations of the image signals with a plurality of references
- G06V30/194—References adjustable by an adaptive method, e.g. learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Health & Medical Sciences (AREA)
- Evolutionary Biology (AREA)
- Databases & Information Systems (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
Description
技术领域technical field
本发明涉及人类视觉仿真技术领域,具体地讲是一种仿真人类低层视觉的目标感知方法。The invention relates to the technical field of human vision simulation, in particular to a target perception method for simulating human low-level vision.
背景技术Background technique
随着信息技术的发展,计算机视觉已经被广泛应用于低层特征检测和描述、模式识别、人工智能推理和机器学习算法等领域。然而,计算机视觉是一种任务驱动型的方法,即需要限定许多条件,并根据实际任务来设计相应的算法,缺乏通用的算法,因而经常会遇到高维非线性特征空间、超大数据量对问题求解和实时处理等问题,使得其研究和应用面临巨大的挑战。With the development of information technology, computer vision has been widely used in low-level feature detection and description, pattern recognition, artificial intelligence reasoning and machine learning algorithms and other fields. However, computer vision is a task-driven method, that is, it needs to limit many conditions and design corresponding algorithms according to actual tasks. It lacks general algorithms, so it often encounters high-dimensional nonlinear feature spaces and large amounts of data to solve problems. And real-time processing and other issues make its research and application face great challenges.
对于人类视觉系统来说,能够在不同环境下高效、可靠地工作,其具有以下优点:具有关注机制、显著性检测以及与此相关的视觉处理中的选择性和目的性;能够从低层视觉处理中利用先验知识,使数据驱动的自底向上处理与自顶向下的知识指导在视觉处理中相互协调配合;上下境信息在视觉处理的各个层次都发挥着重要作用,并且能够综合利用环境中各种模态的信息。但在人类视觉感知机理尚不完全明了的情况下,如何构造具有人类视觉特点的机器视觉仍存在较大困难,若能够仿真人类视觉以实现对目标的感知,则必然会给目标的识别和感知等应用带来重要的影响。For the human visual system to work efficiently and reliably in different environments, it has the following advantages: having attention mechanism, saliency detection and related selectivity and purpose in visual processing; being able to learn from low-level visual processing The use of prior knowledge in visual processing enables data-driven bottom-up processing and top-down knowledge guidance to coordinate with each other in visual processing; contextual information plays an important role in all levels of visual processing, and can comprehensively utilize the environment Information in various modalities. However, when the mechanism of human visual perception is not yet fully understood, there are still great difficulties in how to construct machine vision with human visual characteristics. and other applications have a significant impact.
发明内容Contents of the invention
有鉴于此,本发明要解决的技术问题是,提供一种能够仿真人类视觉,实现对目标场景的快速有效注视的仿真人类低层视觉的目标感知方法。In view of this, the technical problem to be solved by the present invention is to provide a target perception method for simulating human low-level vision that can simulate human vision and realize fast and effective fixation on target scenes.
本发明的技术解决方案是,提供以下步骤的仿真人类低层视觉的目标感知方法,包括以下各步骤:Technical solution of the present invention is, the object perception method of the simulation human low-level vision of following steps is provided, comprises the following steps:
1)通过频域法对目标图像作显著性检测,得到相应的像素显著度图,所述像素显著度图与所述目标图像的像素位置信息一致;1) performing saliency detection on the target image by a frequency domain method to obtain a corresponding pixel saliency map, the pixel saliency map is consistent with the pixel position information of the target image;
2)对所述的像素显著度图中的显著点,依据显著度进行排序;2) sorting the salient points in the pixel saliency map according to the saliency;
3)选取前N个显著点作为注视点,包含这些注视点的最小矩形范围作为注视区域;3) Select the first N salient points as fixation points, and the minimum rectangular range including these fixation points as the fixation area;
4)对所述的注视区域内部像素进行随机采样,并对注视区域外部进行等量的像素随机采样;采样得到的注视区域内部像素作为正样本,注视区域外部像素作为负样本;4) random sampling is carried out to the internal pixels of the fixation area, and random sampling is carried out to an equal amount of pixels outside the fixation area; the internal pixels of the fixation area obtained by sampling are used as positive samples, and the external pixels of the fixation area are used as negative samples;
5)利用支持向量机训练策略,训练得到一个二分类的SVM模型,通过该模型分类所述目标图像的全部像素,将被分为正样本的像素区域作为第一注视目标区。5) Utilize support vector machine training strategy, train the SVM model that obtains a binary classification, classify all pixels of described target image by this model, will be divided into the pixel area of positive sample as the first gaze target area.
采用本发明的方法,与现有技术相比,本发明具有以下优点:通过频域法进行显著性检测,能够快速形成像素显著度图,该图与目标图像的像素位置信息一致,并根据显著度予以排序,将选取的注视点所构成的最小矩形范围作为注视区域进行采样,与外部样本一起进入神经网络,将显著度高的区域作为第一注视目标区,且可建立第一注视目标区的基础上,再次扩大注视范围,形成相应的注视目标区,并与第一注视目标区进行比较,以判断第一注视目标区的结果是否稳定。本发明根据人类视觉注视的过程,通过注视点排序和神经网络模型,来仿真人类视觉,从而实现对目标场景快速有效注视和感知。Using the method of the present invention, compared with the prior art, the present invention has the following advantages: the saliency detection by the frequency domain method can quickly form a pixel saliency map, which is consistent with the pixel position information of the target image, and according to the saliency Sorting the minimum rectangular range formed by the selected fixation points as the fixation area for sampling, entering the neural network together with external samples, taking the area with high salience as the first fixation target area, and the first fixation target area can be established On the basis of , expand the fixation range again to form a corresponding fixation target area, and compare it with the first fixation target area to judge whether the result of the first fixation target area is stable. According to the process of human visual gazing, the present invention simulates human vision through gazing point sorting and a neural network model, thereby realizing fast and effective gazing and perception of target scenes.
作为改进,选取前N+M个显著点作为注视点,依照步骤3)形成注视区域,再经步骤4)和5)得到相应的第二注视目标区;比较第一注视目标区和第二注视目标区的重叠程度,重叠程度大则表明对目标的视觉感知强度大;重叠程度小则表明还未形成足够的对目标的视觉感知强度,继续重复上述过程,直至达到足够的视觉感知强度,最终的注视目标区为上述过程所有注视目标区的叠加。该设计能够加快视觉感知目标的生成与输出,并得到更为稳定的注视目标区,注视的结果更为可靠。As an improvement, select the first N+M salient points as fixation points, form a fixation area according to step 3), and then obtain the corresponding second fixation target area through steps 4) and 5); compare the first fixation target area and the second fixation target area The degree of overlap of the target area, a large degree of overlap indicates that the visual perception intensity of the target is large; a small degree of overlap indicates that the visual perception intensity of the target has not been formed, and the above process is continued until sufficient visual perception intensity is achieved. The fixation target area of is the superposition of all fixation target areas in the above process. This design can speed up the generation and output of visual perception targets, and obtain a more stable fixation target area, and the fixation result is more reliable.
作为改进,获得注视目标区后,在目标图像和像素显著度图中该区域被清零,对更新后的像素显著度图中的显著点,依据显著度再次排序,重复步骤3)、4)和5),得到新的注视目标区,依次获得图像中的多个目标区。这样能够完成对整幅图像的有效信息进行注视识别和读取,提高注视的准确性和完整度。As an improvement, after the fixation target area is obtained, this area is cleared in the target image and the pixel saliency map, and the salient points in the updated pixel saliency map are reordered according to the saliency, and steps 3) and 4) are repeated and 5), to obtain a new fixation target area, and sequentially obtain multiple target areas in the image. In this way, the gaze recognition and reading of the effective information of the entire image can be completed, and the accuracy and completeness of gaze can be improved.
作为改进,所述的频域法是指通过超复数傅立叶变换,将彩色图像中的红、绿、蓝三个分量作为超复数的三个虚部参与傅立叶变换,只保留相位谱信息,经傅立叶反变换获得像素显著度图。该设计用于解决现有技术仅能处理黑白图像识别的问题,有效地针对彩色图像相应地改进了频域法的具体步骤。As an improvement, the frequency-domain method refers to using the hypercomplex Fourier transform to take the three components of red, green and blue in the color image as the three imaginary parts of hypercomplexity to participate in the Fourier transform, only retaining the phase spectrum information, and performing the Fourier transform The inverse transformation obtains a pixel saliency map. The design is used to solve the problem that the prior art can only deal with black and white image recognition, and effectively improves the specific steps of the frequency domain method for color images.
附图说明Description of drawings
图1为本发明仿真人类低层视觉的目标感知方法的流程图。FIG. 1 is a flow chart of the object perception method for simulating human low-level vision in the present invention.
具体实施方式Detailed ways
下面就具体实施例对本发明作进一步说明,但本发明并不仅仅限于这些实施例。The present invention will be further described below with regard to specific examples, but the present invention is not limited only to these examples.
本发明涵盖任何在本发明的精髓和范围上做的替代、修改、等效方法以及方案。为了使公众对本发明有彻底的了解,在以下本发明优选实施例中详细说明了具体的细节,而对本领域技术人员来说没有这些细节的描述也可以完全理解本发明。此外,本发明之附图中为了示意的需要,并没有完全精确地按照实际比例绘制,在此予以说明。The present invention covers any alternatives, modifications, equivalent methods and schemes made on the spirit and scope of the present invention. In order to provide the public with a thorough understanding of the present invention, specific details are set forth in the following preferred embodiments of the present invention, but those skilled in the art can fully understand the present invention without the description of these details. In addition, for the sake of illustration, the drawings of the present invention are not completely drawn according to the actual scale, and are described here.
如图1所示,本发明的仿真人类低层视觉的目标感知方法,包括以下各步骤:As shown in Figure 1, the object perception method of the simulation human low-level vision of the present invention comprises the following steps:
1)通过频域法对目标图像作显著性检测,得到相应的像素显著度图,所述像素显著度图与所述目标图像的像素位置信息一致;1) performing saliency detection on the target image by a frequency domain method to obtain a corresponding pixel saliency map, the pixel saliency map is consistent with the pixel position information of the target image;
2)对所述的像素显著度图中的显著点,依据显著度进行排序;2) sorting the salient points in the pixel saliency map according to the saliency;
3)选取前N个显著点作为注视点,包含这些注视点的最小矩形范围作为注视区域;选取最小矩形范围既能确保采样的精准,也能提高注视目标区的准确性和稳定性;3) Select the first N salient points as fixation points, and the smallest rectangular range containing these fixation points as the fixation area; selecting the smallest rectangular range can not only ensure the accuracy of sampling, but also improve the accuracy and stability of the fixation target area;
4)对所述的注视区域内部像素进行随机采样,并对注视区域外部进行等量的像素随机采样;采样得到的注视区域内部像素作为正样本,注视区域外部像素作为负样本;4) random sampling is carried out to the internal pixels of the fixation area, and random sampling is carried out to an equal amount of pixels outside the fixation area; the internal pixels of the fixation area obtained by sampling are used as positive samples, and the external pixels of the fixation area are used as negative samples;
5)利用支持向量机训练策略,训练得到一个二分类的SVM模型,通过该模型分类所述目标图像的全部像素,将被分为正样本的像素区域作为第一注视目标区。5) Utilize support vector machine training strategy, train the SVM model that obtains a binary classification, classify all pixels of described target image by this model, will be divided into the pixel area of positive sample as the first gaze target area.
对于仿真人类视觉进行感知来说,图像相当于人类视觉所注视的场景,无论场景大小,在视网膜上成像的范围的不变的,因而图像之于机器之于机器视觉也是如此。For the perception of simulating human vision, the image is equivalent to the scene that human vision is looking at. Regardless of the size of the scene, the range of imaging on the retina remains unchanged. Therefore, the same is true of images for machines and machine vision.
通过频域法对目标图像作显著性检测,可采用以下步骤实施:对待目标图像I(i,j)进行二维离散傅里叶变换F[I(i,j)],将图像由空间域转换到频域,得到相位P(u,v)信息:The saliency detection of the target image by the frequency domain method can be implemented by the following steps: perform two-dimensional discrete Fourier transform F[I(i, j)] on the target image I(i, j), transform the image from the spatial domain Convert to the frequency domain to get the phase P(u, v) information:
式中F表示二维离散傅里叶变换,表示相位运算。将相位信息经傅里叶逆变换后,可以在空间域得到显著度图像Sa_Map。In the formula, F represents the two-dimensional discrete Fourier transform, Indicates phase operation. After the phase information is inversely transformed by Fourier, the saliency image Sa_Map can be obtained in the space domain.
Sa_Map(i,j)=|F-1[exp{jP(u,v)}]|2 (2)Sa_Map(i, j)=|F -1 [exp{jP(u, v)}]| 2 (2)
图1中,涉及训练数据、分类模型、结果等均为采用支持向量机(SVM)训练策略相应实施过程。具体实施过程如下:In Figure 1, training data, classification models, results, etc. are all implemented using support vector machine (SVM) training strategies. The specific implementation process is as follows:
设包含l个样本的训练集为输入向量,yk∈{-1,+1}为正负类别标识。SVM首先要用训练集学习建模,目的是在特征空间寻找最优分类超平面,将测试数据尽可能正确地分类。考虑一般情况,训练集为非线性可分时,先选择一个高斯径向基核函数Suppose the training set contains l samples is the input vector, and y k ∈ {-1, +1} is the positive and negative category identification. SVM first uses the training set to learn modeling, the purpose is to find the optimal classification hyperplane in the feature space, and classify the test data as correctly as possible. Considering the general situation, when the training set is nonlinearly separable, first choose a Gaussian radial basis kernel function
K(x,xi)=exp{-q||x-xi||2} (3)K(x, x i )=exp{-q||xx i || 2 } (3)
将训练集数据xi映射到一个高维线性特征空间中构造最优分类超平面。其中q为径向基核函数参数,则分类器的判别函数为Map the training set data xi to a high-dimensional linear feature space to construct the optimal classification hyperplane. Where q is the parameter of the radial basis kernel function, and the discriminant function of the classifier is
训练过程是已知和q等条件下,利用二次规划求解方法获得(4)式中的b*,αi *和支持向量(SV)作为训练得到的SVM模型;测试过程则是利用该SVM模型,将未知的数据x代入(4)式,获得其预测类别。The training process is known and q, etc., use the quadratic programming method to obtain the b * , α i * and support vector (SV) in the formula (4) as the SVM model obtained by training; the test process is to use the SVM model to convert the unknown Data x is substituted into formula (4) to obtain its predicted category.
SVM利用核函数技巧避免了传统学习算法面临的维数灾难问题。基于结构风险最小化原则,其分类性能只由少量支持向量(SV)决定,具有好的泛化性能。在实际问题中,有利于利用先验知识挑选少量样本,经SVM学习来构造分类器。其克服了传统学习算法基于经验风险最小化原则,当样本数趋向于无穷大时性能才有理论上的保证的缺陷;通过求解二次规划问题,能够避免传统神经网络算法构建网络的经验性和容易陷入局部极小解等缺点;适合分割复杂、难以定量描述的图像目标。SVM uses the technique of kernel function to avoid the curse of dimensionality problem faced by traditional learning algorithms. Based on the principle of structural risk minimization, its classification performance is only determined by a small number of support vectors (SV), and it has good generalization performance. In practical problems, it is beneficial to use prior knowledge to select a small number of samples and construct a classifier through SVM learning. It overcomes the defect that the traditional learning algorithm is based on the principle of empirical risk minimization, and the performance can only be guaranteed theoretically when the number of samples tends to infinity; by solving the quadratic programming problem, it can avoid the empirical and easy construction of the traditional neural network algorithm Falling into local minimum solutions and other shortcomings; it is suitable for image targets that are complex to segment and difficult to describe quantitatively.
为了优化本发明,则需要判断得到的第一注视目标区是否稳定,框图中则体现为判断是否存在稳定的输出。因此需要形成进一步的目标区:In order to optimize the present invention, it is necessary to judge whether the obtained first fixation target area is stable, which is reflected in the block diagram to judge whether there is a stable output. Therefore further target areas need to be formed:
选取前N+M个显著点作为注视点,依照步骤3)形成注视区域,再经步骤4)和5)得到相应的第二注视目标区;比较第一注视目标区和第二注视目标区的重叠程度,重叠程度大则表明对目标的视觉感知强度大;重叠程度小则表明还未形成足够的对目标的视觉感知强度,继续重复上述过程,直至达到足够的视觉感知强度,最终的注视目标区为上述过程所有注视目标区的叠加。Select the first N+M salient points as fixation points, form a fixation area according to step 3), and then obtain the corresponding second fixation target area through steps 4) and 5); compare the first fixation target area and the second fixation target area The degree of overlap, a large degree of overlap indicates that the visual perception intensity of the target is large; a small degree of overlap indicates that the visual perception intensity of the target has not been formed, and the above process continues to be repeated until sufficient visual perception intensity is achieved, and the final gaze target Region is the superposition of all fixation target regions in the above process.
获得注视目标区后,在目标图像和像素显著度图中该区域被清零,对更新后的像素显著度图中的显著点,依据显著度再次排序,重复步骤3)、4)和5),得到新的注视目标区,依次获得图像中的多个目标区。这样便可从图中分割出所有有效注视区域的信息,构建了模拟人类视觉的机器视觉。After the fixation target area is obtained, the area is cleared in the target image and the pixel saliency map, and the salient points in the updated pixel saliency map are sorted again according to the saliency, and steps 3), 4) and 5) are repeated , to obtain a new fixation target area, and to sequentially obtain multiple target areas in the image. In this way, the information of all effective gaze areas can be segmented from the image, and a machine vision that simulates human vision can be constructed.
所述的频域法是指通过超复数傅立叶变换,将彩色图像中的红、绿、蓝三个分量作为超复数的三个虚部参与傅立叶变换,只保留相位谱信息,经傅立叶反变换获得像素显著度图。该设计用于解决现有技术仅能处理黑白图像识别的问题,有效地针对彩色图像相应地改进了频域法的具体步骤。超复数由四个部分组成,表示为The frequency domain method refers to the use of the hypercomplex Fourier transform to use the red, green and blue components of the color image as the three imaginary parts of the hypercomplex to participate in the Fourier transform, only retaining the phase spectrum information, which is obtained by inverse Fourier transform Pixel saliency map. The design is used to solve the problem that the prior art can only deal with black and white image recognition, and effectively improves the specific steps of the frequency domain method for color images. A hypercomplex number consists of four parts, denoted as
q=a+bi+cj+dk (5)q=a+bi+cj+dk (5)
其中a,b,c,d都是实数,i,j,k都是虚数单位,且具有以下性质:i2=j2=k2=ijk=-1,ij=-ji=k,ki=-ik=j,jk=-kj=i。Where a, b, c, d are all real numbers, i, j, k are all imaginary units, and have the following properties: i 2 =j 2 =k 2 =ijk=-1, ij=-ji=k, ki= -ik=j, jk=-kj=i.
彩色图像的RGB模型可以描述为没有实部的纯超复数:The RGB model of a color image can be described as a pure hypercomplex number with no real part:
f=R(m,n)i+G(m,n)j+B(m,n)k (6)f=R(m,n)i+G(m,n)j+B(m,n)k (6)
其中R(m,n),G(m,n),B(m,n)分别表示图像红绿蓝三个分量,m,n为像素坐标。若q=f,则a=0,b=R(m,n),c=G(m,n),d=B(m,n)。对彩色矢量可按照式(6)进行超复数傅里叶变换:Among them, R(m, n), G(m, n), and B(m, n) represent the three components of image red, green and blue, respectively, and m and n are pixel coordinates. If q=f, then a=0, b=R(m,n), c=G(m,n), d=B(m,n). The hypercomplex Fourier transform can be performed on the color vector according to formula (6):
FR(v,u)=(real(fft2(a))+μ·imag(fft2(a)))+F R (v,u)=(real(fft2(a))+μ·imag(fft2(a)))+
i(real(fft2(b))+μ·imag(fft2(b)))+ (7)i(real(fft2(b))+μ imag(fft2(b)))+ (7)
j(real(fft2(c))+μ·imag(fft2(c)))+j(real(fft2(c))+μ·imag(fft2(c)))+
k(real(fft(d))+μ·imag(fft2(d)))k(real(fft(d))+μ·imag(fft2(d)))
其中,fft2()表示传统二维傅里叶变换,real()表示取实部,imag()表示取虚部。Among them, fft2() represents the traditional two-dimensional Fourier transform, real() represents the real part, and imag() represents the imaginary part.
为单位虚向量。此处,只需取FR(v,u)的相位谱P(f): is a unit virtual vector. Here, just take the phase spectrum P(f) of FR ( v, u):
令:A=ejP(f) (9)Order: A=e jP(f) (9)
利用传统二维快速傅里叶逆变换(ifft2)组合可以得到超复数傅里叶逆变换,如式(10):Using the traditional two-dimensional inverse fast Fourier transform (ifft2) combination can obtain the hypercomplex inverse Fourier transform, as shown in formula (10):
F-R(v,u)=(real(ifft2(A))+μ·imag(ifft2(A)))+F -R (v,u)=(real(ifft2(A))+μ·imag(ifft2(A)))+
i(real(ifft2(B))+μ·imag(ifft2(B)))+ (10)i(real(ifft2(B))+μ imag(ifft2(B)))+ (10)
j(real(ifft2(C))+μ·imag(ifft2(C)))+j(real(ifft2(C))+μ·imag(ifft2(C)))+
k(real(ifft2(D))+μ·imag(ifft2(D)))k(real(ifft2(D))+μ·imag(ifft2(D)))
其中,B=fft2(b),C=fft2(c),D=fft2(d)。Wherein, B=fft2(b), C=fft2(c), D=fft2(d).
Sa_Map(m,n)=real(F-R(v,u)) (11)Sa_Map(m,n)=real(F- R (v,u)) (11)
即为求得的显著图。由于彩色像素在数据处理前后的整体性得到了保持,从而避免了由于矢量分量的变换或交换引起的色彩失真。is the obtained saliency map. Since the integrity of color pixels before and after data processing is maintained, color distortion caused by transformation or exchange of vector components is avoided.
以上仅就本发明较佳的实施例作了说明,但不能理解为是对权利要求的限制。本发明不仅局限于以上实施例,其具体结构允许有变化。总之,凡在本发明独立权利要求的保护范围内所作的各种变化均在本发明的保护范围内。The above is only an illustration of the preferred embodiments of the present invention, but should not be construed as a limitation on the claims. The present invention is not limited to the above embodiments, and its specific structure is allowed to vary. In a word, all kinds of changes made within the protection scope of the independent claims of the present invention are within the protection scope of the present invention.
Claims (4)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201510377158.5A CN105005788B (en) | 2015-06-25 | 2015-06-25 | The target apperception method of simulated human Low Level Vision |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201510377158.5A CN105005788B (en) | 2015-06-25 | 2015-06-25 | The target apperception method of simulated human Low Level Vision |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN105005788A true CN105005788A (en) | 2015-10-28 |
| CN105005788B CN105005788B (en) | 2018-08-28 |
Family
ID=54378453
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201510377158.5A Expired - Fee Related CN105005788B (en) | 2015-06-25 | 2015-06-25 | The target apperception method of simulated human Low Level Vision |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN105005788B (en) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN107392120A (en) * | 2017-07-06 | 2017-11-24 | 电子科技大学 | A kind of notice intelligence direct method based on sight estimation |
| CN112418296A (en) * | 2020-11-18 | 2021-02-26 | 中国科学院上海微系统与信息技术研究所 | A bionic binocular target recognition and tracking method based on human visual attention mechanism |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US7940985B2 (en) * | 2007-06-06 | 2011-05-10 | Microsoft Corporation | Salient object detection |
| CN104574335A (en) * | 2015-01-14 | 2015-04-29 | 西安电子科技大学 | Infrared and visible image fusion method based on saliency map and interest point convex hulls |
-
2015
- 2015-06-25 CN CN201510377158.5A patent/CN105005788B/en not_active Expired - Fee Related
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US7940985B2 (en) * | 2007-06-06 | 2011-05-10 | Microsoft Corporation | Salient object detection |
| CN104574335A (en) * | 2015-01-14 | 2015-04-29 | 西安电子科技大学 | Infrared and visible image fusion method based on saliency map and interest point convex hulls |
Non-Patent Citations (4)
| Title |
|---|
| XIAODI HOU: "Saliency Detection: A Spectral Residual Approach", 《COMPUTER VISION AND PATTERN RECOGNITION(CVPR),2007IEEE CONFERENCE ON》 * |
| 侯庆岑: "模拟人类视觉的自动图像分割技术研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
| 杨军 等: "基于全局颜色对比的显著性目标检测", 《计算机应用研究》 * |
| 郭涛 等: "基于视觉注意的SVM彩色图像分割方法", 《计算机工程与应用》 * |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN107392120A (en) * | 2017-07-06 | 2017-11-24 | 电子科技大学 | A kind of notice intelligence direct method based on sight estimation |
| CN107392120B (en) * | 2017-07-06 | 2020-04-14 | 电子科技大学 | An attention-intelligent supervision method based on line-of-sight estimation |
| CN112418296A (en) * | 2020-11-18 | 2021-02-26 | 中国科学院上海微系统与信息技术研究所 | A bionic binocular target recognition and tracking method based on human visual attention mechanism |
| CN112418296B (en) * | 2020-11-18 | 2024-04-02 | 中国科学院上海微系统与信息技术研究所 | Bionic binocular target identification and tracking method based on human eye visual attention mechanism |
Also Published As
| Publication number | Publication date |
|---|---|
| CN105005788B (en) | 2018-08-28 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN113066002B (en) | Method for generating adversarial samples, method, device and equipment for training neural network | |
| Eslami et al. | Attend, infer, repeat: Fast scene understanding with generative models | |
| CN107609541B (en) | Human body posture estimation method based on deformable convolution neural network | |
| CN104992183B (en) | The automatic testing method of well-marked target in natural scene | |
| Chen et al. | Diagnose like a pathologist: Weakly-supervised pathologist-tree network for slide-level immunohistochemical scoring | |
| CN113223037B (en) | Unsupervised semantic segmentation method and unsupervised semantic segmentation system for large-scale data | |
| Pan et al. | A learning-based positive feedback approach in salient object detection | |
| CN104992452A (en) | Flight object automatic tracking method based on thermal imaging video | |
| CN110348475A (en) | It is a kind of based on spatial alternation to resisting sample Enhancement Method and model | |
| CN111723875B (en) | Refocusing Method of SAR 3D Rotating Ship Target Based on CV-RefocusNet | |
| Rizvi et al. | Revolutionizing agriculture: Machine and deep learning solutions for enhanced crop quality and weed control | |
| Xue et al. | Region comparison network for interpretable few-shot image classification | |
| CN107247952B (en) | Deep supervision-based visual saliency detection method for cyclic convolution neural network | |
| CN114708462A (en) | Detection model generation method, system, device and storage medium for multi-data training | |
| Khellal et al. | Pedestrian classification and detection in far infrared images | |
| CN116433704A (en) | A cell nucleus segmentation method and related equipment based on center points | |
| Chen et al. | Multi-dimensional color image recognition and mining based on feature mining algorithm | |
| CN104933691A (en) | Image fusion method based on phase spectrum visual saliency detection | |
| CN104933435B (en) | Machine vision construction method based on simulation human vision | |
| CN113792764B (en) | Sample expansion method, system, storage medium and electronic equipment | |
| CN105005788B (en) | The target apperception method of simulated human Low Level Vision | |
| Zhang et al. | A survey of few-shot learning and its application in industrial object detection tasks | |
| CN104933725B (en) | Simulate the image partition method of human vision | |
| Chen | Evaluation technology of classroom students’ learning state based on deep learning | |
| Yang et al. | Adaptive multi-class correlation filters |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| C06 | Publication | ||
| PB01 | Publication | ||
| C10 | Entry into substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant | ||
| CF01 | Termination of patent right due to non-payment of annual fee | ||
| CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20180828 |