CN118072101A - Noisy label image learning method and system based on balanced selection and contrastive learning - Google Patents
Noisy label image learning method and system based on balanced selection and contrastive learning Download PDFInfo
- Publication number
- CN118072101A CN118072101A CN202410281796.6A CN202410281796A CN118072101A CN 118072101 A CN118072101 A CN 118072101A CN 202410281796 A CN202410281796 A CN 202410281796A CN 118072101 A CN118072101 A CN 118072101A
- Authority
- CN
- China
- Prior art keywords
- samples
- true
- training
- marked
- sequence
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/0895—Weakly supervised learning, e.g. semi-supervised or self-supervised learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/0985—Hyperparameter optimisation; Meta-learning; Learning-to-learn
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Software Systems (AREA)
- Multimedia (AREA)
- Medical Informatics (AREA)
- Databases & Information Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Image Analysis (AREA)
Abstract
本发明提供了一种基于均衡选择及对比学习的含噪声标签图像学习方法与系统,涉及人工智能及计算机视觉领域。学习方法包括如下步骤:利用联合损失函数,基于原始数据集样本,对编号为m={1,2}的DNN模型进行若干轮次的预训练;在历史序列Sm中按照时间顺序,将含干净标签样本的索引标记为True,其他索引标记为False,将标记为True的样本放入子集将剩余样本移除标签后放入子集基于子集子集对编号为m={1,2}的DNN模型进行鲁棒训练;在全部轮次的预训练结束后,在历史序列Sm中最后连续个结果中,挑选包含被标记为True的样本数目最多的那组序列,将该序列中的被标记为True的样本放入基准集合Dc;重新初始化编号为m={1,2}的DNN模型,重复上述步骤,直到达到预设的训练总轮次。
The present invention provides a method and system for learning noisy labeled images based on balanced selection and contrastive learning, which relates to the fields of artificial intelligence and computer vision. The learning method includes the following steps: using a joint loss function, based on the original data set samples, pre-training a DNN model numbered m = {1,2} for several rounds; in the historical sequence S m , in chronological order, marking the index of the sample containing the clean label as True, marking the other indexes as False, and putting the samples marked as True into the subset Remove labels from the remaining samples and put them into the subset Subset-based Subset Robust training is performed on the DNN model numbered m = {1,2}; after all rounds of pre-training are completed, the last consecutive Among the results, select the sequence with the largest number of samples marked as True, and put the samples marked as True in the sequence into the benchmark set D c ; reinitialize the DNN model numbered m = {1, 2}, and repeat the above steps until the preset total number of training rounds is reached.
Description
技术领域Technical Field
本发明涉及人工智能及计算机视觉领域,尤其涉及一种基于均衡选择及对比学习的含噪声标签图像学习方法与系统。The present invention relates to the fields of artificial intelligence and computer vision, and in particular to a noisy label image learning method and system based on balanced selection and contrast learning.
背景技术Background technique
过去数十年,深度神经网络(Deep Neural Networks,DNN)在图像处理领域取得巨大成功。其发展受到了大规模具有良好注释的数据集的影响,然而这类数据集的收集需要耗费巨大的人力物力,阻碍了DNN的进一步应用。而弱监督和半监督学习由于对模型的标注质量要求较低,受到巨大关注,其使用的数据集通常含有大量不准确标签(噪声标签),而传统训练方法使模型易过拟合噪声标签,导致测试性能和泛化性能的下降。In the past few decades, deep neural networks (DNNs) have achieved great success in the field of image processing. Its development has been influenced by large-scale well-annotated datasets. However, the collection of such datasets requires huge manpower and material resources, which hinders the further application of DNNs. Weakly supervised and semi-supervised learning have received great attention due to their low requirements on the quality of model annotation. The datasets they use usually contain a large number of inaccurate labels (noise labels), and traditional training methods make the model prone to overfitting the noisy labels, resulting in a decrease in test performance and generalization performance.
存在现有技术,基于小损失准则,使用两个模型交替为彼此挑选含干净标签的图像样本进行训练,但是其性能较差。There are existing technologies that use two models to alternately select image samples with clean labels for each other for training based on the small loss criterion, but their performance is poor.
此外,还存在现有技术,将引入混合高斯模型基于样本的训练损失进行建模,从而识别噪声标签样本,随后利用半监督方法训练取得了一定的成果,但是该方法在高噪声比率的数据集上表现交叉。In addition, there is an existing technology that introduces a mixed Gaussian model to model based on sample training loss to identify noise label samples, and then uses a semi-supervised method to train and achieve certain results, but this method performs poorly on data sets with high noise ratios.
还存在现有技术,首先通过滑窗方式收集一个尺寸较小但噪声比率较低的干净基准集,然后在第二阶段基于该基准集合进行模型的性能优化,取得了一定的进步,但是其忽略了基准集合的类别不均衡问题,导致DNN在非对称噪声场景下的性能下降。There is also an existing technology that first collects a clean benchmark set with a smaller size but lower noise ratio through a sliding window method, and then optimizes the performance of the model based on the benchmark set in the second stage. It has made certain progress, but it ignores the category imbalance problem of the benchmark set, resulting in a decrease in the performance of DNN in asymmetric noise scenarios.
发明内容Summary of the invention
发明目的:本发明提出一种基于均衡选择及对比学习的含噪声标签图像学习方法与系统,旨在有效解决现有技术存在的上述问题。Purpose of the invention: The present invention proposes a noisy label image learning method and system based on balanced selection and contrastive learning, aiming to effectively solve the above-mentioned problems existing in the prior art.
第一方面,提出一种基于均衡选择及对比学习的含噪声标签图像学习方法,步骤如下:First, a noisy label image learning method based on balanced selection and contrastive learning is proposed. The steps are as follows:
S1、利用联合损失函数,基于原始数据集样本,对编号为m={1,2}的DNN模型进行若干轮次的预训练;S1. Using the joint loss function, the DNN model numbered m = {1, 2} is pre-trained for several rounds based on the original dataset samples.
在历史序列Sm中按照时间顺序,将含干净标签样本的索引标记为True,其他索引标记为False,将标记为True的样本放入子集将剩余样本移除标签后放入子集/> In the historical sequence Sm , in chronological order, the index containing the clean label sample is marked as True, and the other indexes are marked as False. The samples marked as True are put into the subset Remove labels from the remaining samples and put them into the subset/>
S2、基于所述子集子集/>对编号为m={1,2}的DNN模型进行鲁棒训练;S2. Based on the subset Subset/> Perform robust training on the DNN model numbered m={1,2};
S3、在全部轮次的预训练结束后,在历史序列Sm中最后连续个结果中,挑选包含被标记为True的样本数目最多的那组序列,将该序列中的被标记为True的样本放入基准集合Dc;S3, after all rounds of pre - training are completed, the last consecutive Among the results, select the sequence with the largest number of samples marked as True, and put the samples marked as True in the sequence into the benchmark set D c ;
S4、重新初始化编号为m={1,2}的DNN模型,重复步骤S1至S3,直到达到预设的训练总轮次。S4. Reinitialize the DNN model numbered m={1,2} and repeat steps S1 to S3 until the preset total number of training rounds is reached.
在第一方面进一步的实施例中,步骤S1中所述原始数据集样本的表达式如下:In a further embodiment of the first aspect, the original data set sample in step S1 The expression is as follows:
式中,n是数据集样本个数;是图像xi的观测标签,c表示数据集包含的类别数。In the formula, n is the number of samples in the data set; is the observed label of image xi , and c represents the number of categories contained in the dataset.
在第一方面进一步的实施例中,步骤S1中所述联合损失函数Ljoint的表达式如下:In a further embodiment of the first aspect, the expression of the joint loss function L joint in step S1 is as follows:
式中,xi表示输入图像;n表示数据集样本个数;p(xi)表示输入特征xi的预测输出值。In the formula, xi represents the input image; n represents the number of samples in the dataset; and p( xi ) represents the predicted output value of the input feature xi .
在第一方面进一步的实施例中,对编号为m={1,2}的DNN模型进行若干轮次的预训练,进一步包括:In a further embodiment of the first aspect, performing several rounds of pre-training on the DNN model numbered m={1,2} further includes:
计算初始噪声数据集中所有样本在DNN模型下的JSD损失di:Calculate the JSD loss d i of all samples in the initial noise dataset under the DNN model:
式中,p1(xi)表示编号m=1的DNN模型基于图像xi的输出值;p2(xi)表示编号m=2的DNN模型基于图像xi的输出值;Wherein, p 1 ( xi ) represents the output value of the DNN model with number m=1 based on the image x ; p 2 ( xi ) represents the output value of the DNN model with number m=2 based on the image x ;
KL(*)表示Kullback-Leibler函数,如下:KL(*) represents the Kullback-Leibler function, as follows:
式中,m和n表示输入参数。。Where m and n represent input parameters.
在第一方面进一步的实施例中,将标记为True的样本放入子集所述子集/>的表达式如下:In a further embodiment of the first aspect, samples marked as True are placed into a subset The subset/> The expression is as follows:
式中,xi表示输入图像;表示输入图像xi的观测标签;Sm表示历史序列;/>表示第i个样本在第t个epoch的权重值;e=t-K+1表示第t-K+1个epoch;/>表示第i个样本在第e个epoch的权重;/>表示第i个样本在第j个epoch的权重;/>表示所有样本在第t-j个epoch时基于第m个网络产生的权重向量。Where, xi represents the input image; represents the observed label of the input image x i ; S m represents the historical sequence; /> represents the weight value of the i-th sample at the t-th epoch; e=t-K+1 represents the t-K+1-th epoch;/> Indicates the weight of the i-th sample in the e-th epoch; /> Indicates the weight of the i-th sample in the j-th epoch; /> Represents the weight vector generated by all samples based on the m-th network at the tj-th epoch.
在第一方面进一步的实施例中,所述历史序列 In a further embodiment of the first aspect, the historical sequence
式中,为数据集/>中所有样本在第t个epoch时,基于编号为m的模型的标记结果。In the formula, For the dataset/> The labeling results of all samples in the tth epoch based on the model numbered m.
在第一方面进一步的实施例中,在步骤S1中,对于编号为m=1的模型:In a further embodiment of the first aspect, in step S1, for the model numbered m=1:
基于所述JSD损失di,对每个类别包含的样本按照其对应的散度值进行排序,并挑选相同数量的样本作为含干净标签的样本,在历史序列S1中按照时间顺序,将含干净标签样本的索引标记为True,其他索引标记为False;Based on the JSD loss d i , the samples contained in each category are sorted according to their corresponding divergence values, and the same number of samples are selected as samples with clean labels. In the historical sequence S 1 , the indexes of the samples with clean labels are marked as True and the other indexes are marked as False in chronological order;
初始化集合和/>为空,选择序列S1中连续K个epoch均被标记为True的样本放入干净样本集合/>剩余样本移除标签后放入/> Initializing a Collection and/> Empty, select samples in sequence S1 that are marked as True for K consecutive epochs and put them into the clean sample set/> Remove the labels from the remaining samples and place them in />
在第一方面进一步的实施例中,在步骤S1中,对于编号为m=2的模型:In a further embodiment of the first aspect, in step S1, for the model numbered m=2:
基于所述JSD损失di,对每个类别包含的样本按照其对应的散度值进行排序,并挑选相同数量的样本作为含干净标签的样本,在历史序列S2中按照时间顺序,将含干净标签样本的索引标记为True,其他索引标记为False;Based on the JSD loss d i , the samples contained in each category are sorted according to their corresponding divergence values, and the same number of samples are selected as samples with clean labels. In the historical sequence S 2 , the indexes of the samples with clean labels are marked as True and the other indexes are marked as False in chronological order;
初始化集合和/>为空,选择序列S2中连续K个epoch均被标记为True的样本放入干净样本集合/>剩余样本移除标签后放入/> Initializing a Collection and/> is empty, select samples in sequence S 2 that are marked as True for K consecutive epochs and put them into the clean sample set/> Remove the labels from the remaining samples and place them in />
本发明的第二个方面,提出一种含噪声标签图像学习系统,该系统包括预训练模块、鲁棒训练模块、挑选模块、重复执行模块。A second aspect of the present invention provides a noisy label image learning system, which includes a pre-training module, a robust training module, a selection module, and a repeated execution module.
预训练模块利用联合损失函数,基于原始数据集样本,对编号为m={1,2}的DNN模型进行若干轮次的预训练;在历史序列Sm中按照时间顺序,将含干净标签样本的索引标记为True,其他索引标记为False,将标记为True的样本放入子集将剩余样本移除标签后放入子集/> The pre-training module uses the joint loss function to perform several rounds of pre-training on the DNN model numbered m = {1,2} based on the original dataset samples; in the historical sequence S m , the index of the sample containing the clean label is marked as True, and the other indexes are marked as False in chronological order, and the samples marked as True are put into the subset Remove labels from the remaining samples and put them into the subset/>
鲁棒训练模块基于所述子集子集/>对编号为m={1,2}的DNN模型进行鲁棒训练。The robust training module is based on the subset Subset/> Robust training is performed on the DNN model numbered m={1,2}.
挑选模块用于在全部轮次的预训练结束后,在历史序列Sm中最后连续个结果中,挑选包含被标记为True的样本数目最多的那组序列,将该序列中的被标记为True的样本放入基准集合Dc。The selection module is used to select the last consecutive Among the results, the sequence with the largest number of samples marked as True is selected, and the samples marked as True in the sequence are put into the reference set D c .
重复执行模块用于重新初始化编号为m={1,2}的DNN模型,反馈至预训练模块、鲁棒训练模块、挑选模块,直到达到预设的训练总轮次。The repeated execution module is used to reinitialize the DNN model numbered m={1,2}, and feed back to the pre-training module, the robust training module, and the selection module until the preset total number of training rounds is reached.
本发明的第三个方面,提出一种计算机可读存储介质,存储介质中存储有至少一个可执行指令,所述可执行指令在电子设备上运行时,使得电子设备执行如第一方面所述的基于均衡选择及对比学习的含噪声标签图像学习方法。According to a third aspect of the present invention, a computer-readable storage medium is provided, wherein at least one executable instruction is stored in the storage medium. When the executable instruction is executed on an electronic device, the electronic device executes the noisy label image learning method based on balanced selection and contrast learning as described in the first aspect.
有益效果:本发明提出了一种基于均衡选择及对比学习的含噪声标签图像学习方法与系统,该方法提出新的均衡选择策略,以收集一个类别均衡且噪声比率极低的干净子集,随后利用对比学习技术进一步提高模型特征提取能力和测试性能,使模型适用于各类复杂噪声场景,如对称噪声、非对称噪声、实例相关噪声及混合噪声等。利用本发明提出的含噪声标签图像学习方法,与常见现有技术相比,在精度方面具有显著提升。Beneficial effects: The present invention proposes a noisy label image learning method and system based on balanced selection and contrastive learning. The method proposes a new balanced selection strategy to collect a clean subset with balanced categories and extremely low noise ratio, and then uses contrastive learning technology to further improve the model feature extraction ability and test performance, so that the model is suitable for various complex noise scenes, such as symmetrical noise, asymmetrical noise, instance-related noise and mixed noise. Compared with the common existing technology, the noisy label image learning method proposed by the present invention has a significant improvement in accuracy.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
图1为基于均衡选择及对比学习的含噪声标签图像学习方法的整体流程。Figure 1 shows the overall process of the noisy label image learning method based on balanced selection and contrastive learning.
图2为均衡选择的具体流程。Figure 2 shows the specific process of equilibrium selection.
具体实施方式Detailed ways
在下文的描述中,给出了大量具体的细节以便提供对本发明更为彻底的理解。然而,对于本领域技术人员而言显而易见的是,本发明可以无需一个或多个这些细节而得以实施。在其他例子中,为了避免与本发明发生混淆,对于本领域公知的一些技术特征未进行描述。In the following description, a large number of specific details are provided to provide a more thorough understanding of the present invention. However, it is apparent to those skilled in the art that the present invention can be implemented without one or more of these details. In other examples, in order to avoid confusion with the present invention, some technical features well known in the art are not described.
实施例1:Embodiment 1:
针对现有技术的不足之处,本发明提出一种均衡选择及对比学习的含噪声标签图像学习方法,该方法针对提出新的均衡选择策略,以收集一个类别均衡且噪声比率极低的干净子集,随后利用对比学习技术进一步提高模型特征提取能力和测试性能,使模型适用于各类复杂噪声场景,如对称噪声、非对称噪声、实例相关噪声及混合噪声等。本发明为了实现上述目的,采用的技术方法具体如下:In view of the shortcomings of the prior art, the present invention proposes a noisy label image learning method of balanced selection and contrastive learning. The method proposes a new balanced selection strategy to collect a clean subset with balanced categories and extremely low noise ratio, and then uses contrastive learning technology to further improve the model feature extraction ability and test performance, so that the model is suitable for various complex noise scenes, such as symmetrical noise, asymmetrical noise, instance-related noise and mixed noise. In order to achieve the above purpose, the technical methods adopted by the present invention are as follows:
Step 1.1、为两个结构完全一致但是初始化参数不一致的DNN模型(模型编号为m={1,2}),使用联合损失函数(Joint loss function)对给定的含噪声标签数据集进行共Tw轮次(epochs)的温和(warm-up)训练。在每个epoch结束后,计算所有训练样本在两个模型上的Jensen–Shannon divergence,并利用小损失准则,对每个类别包含的样本按照其对应的散度值进行排序,并挑选相同数量的样本作为含干净标签的样本,在历史序列S1和S2中按照时间顺序,将含干净标签样本的索引标记为True值,其他索引标记为False值;Step 1.1, for two DNN models with the same structure but different initialization parameters (model number is m = {1, 2}), use the joint loss function to calculate the given noisy label data set Perform warm-up training for a total of T w epochs. After each epoch, calculate the Jensen–Shannon divergence of all training samples on the two models, and use the small loss criterion to sort the samples contained in each category according to their corresponding divergence values, and select the same number of samples as samples with clean labels. In the historical sequences S 1 and S 2 , mark the indexes of samples with clean labels as True values and other indexes as False values in chronological order;
Step 1.2、对于编号为m=1的模型,在每个epoch的开始,首先计算所有训练样本在两个模型上的Jensen–Shannon divergence,并利用小损失准则,对每个类别包含的样本按照其对应的散度值进行排序,并挑选相同数量的样本作为含干净标签的样本,在历史序列S1中按照时间顺序,将含干净标签样本的索引标记为True值,其他索引标记为False值;Step 1.2: For the model with number m=1, at the beginning of each epoch, first calculate the Jensen–Shannon divergence of all training samples on the two models, and use the small loss criterion to sort the samples contained in each category according to their corresponding divergence values, and select the same number of samples as samples with clean labels. In the historical sequence S1 , in chronological order, mark the index of the sample with clean label as True value, and mark other indexes as False value;
然后,初始化集合和/>为空,选择序列S1中连续K个epoch均被标记为True的样本放入干净样本集合/>剩余样本移除标签后放入/> Then, initialize the collection and/> Empty, select samples in sequence S1 that are marked as True for K consecutive epochs and put them into the clean sample set/> Remove the labels from the remaining samples and place them in />
Step 1.3、使用SSL技术和对比学习技术,基于Step1.2划分出的两个子集和对编号为m=1的模型进行鲁棒训练。Step 1.3: Use SSL technology and contrastive learning technology to divide the two subsets based on Step 1.2. and Perform robust training on the model numbered m=1.
Step 1.4、对于编号为m=2的模型,在每个epoch的开始,首先计算所有训练样本在两个模型上的Jensen–Shannon divergence,并利用小损失准则,对每个类别包含的样本按照其对应的散度值进行排序,并挑选相同数量的样本作为含干净标签的样本,在历史序列S2中按照时间顺序,将含干净标签样本的索引标记为True值,其他索引标记为False值;然后,初始化集合和/>为空,选择序列S2中连续K个epoch均被标记为True的样本放入干净样本集合/>剩余样本移除标签后放入/> Step 1.4, for the model numbered m = 2, at the beginning of each epoch, first calculate the Jensen–Shannon divergence of all training samples on the two models, and use the small loss criterion to sort the samples contained in each category according to their corresponding divergence values, and select the same number of samples as samples with clean labels. In the historical sequence S 2 , in chronological order, mark the index of the sample with clean labels as True value, and mark other indexes as False value; then, initialize the set and/> is empty, select samples in sequence S 2 that are marked as True for K consecutive epochs and put them into the clean sample set/> Remove the labels from the remaining samples and place them in />
Step 1.5、使用SSL技术和对比学习技术,基于Step1.2划分出的两个子集和对编号为m=2的模型进行鲁棒训练。Step 1.5: Use SSL technology and contrastive learning technology to divide the two subsets based on Step 1.2 and Robust training is performed on the model numbered m=2.
Step 1.6如果没有达到预设的训练总轮次Ttot,则跳转到step 1.2继续进行训练。Step 1.6: If the preset total number of training rounds T tot is not reached, jump to step 1.2 to continue training.
Step 1.7步骤1的训练全部结束后,在历史序列S1中最后连续个结果中,挑选包含被标记为True的样本数目最多的那组序列,将该序列中的被标记为True的样本放入基准集合Dc。Step 1.7 After all the training in step 1 is completed, the last continuous Among the results, the sequence with the largest number of samples marked as True is selected, and the samples marked as True in the sequence are put into the reference set D c .
Step 1.8、重新初始化模型编号为m={1,2}的DNN模型,同样使用联合损失函数(Joint loss function)基于给定的含噪声标签数据集进行共Tw轮次(epochs)的温和(warm-up)训练。Step 1.8, re-initialize the DNN model with model number m = {1,2}, and also use the joint loss function based on the given noisy label dataset Perform warm-up training for a total of T w epochs.
Step 1.9、对于编号为m=1的模型,在每个epoch的开始,首先计算所有训练样本在两个模型上的Jensen–Shannon divergence,并利用小损失准则,对每个类别包含的样本按照其对应的散度值进行排序,并挑选相同数量的样本作为含干净标签的样本,将其索引标记为True值,此外,对于被包含在在基准集合Dc中样本,其索引也被标记为True,所有剩余样本的索引被标记为False值;然后,初始化集合和/>为空,将被标记为True的样本放入干净样本集合/>剩余False样本移除标签后放入/> Step 1.9, for the model numbered m=1, at the beginning of each epoch, first calculate the Jensen–Shannon divergence of all training samples on the two models, and use the small loss criterion to sort the samples contained in each category according to their corresponding divergence values, and select the same number of samples as samples with clean labels, and mark their indexes as True values. In addition, for the samples included in the benchmark set D c , their indexes are also marked as True, and the indexes of all remaining samples are marked as False values; then, initialize the set and/> Empty, the samples marked as True are put into the clean sample set/> The remaining False samples are removed from the label and put into />
Step 1.10、使用SSL技术和对比学习技术,基于Step1.9划分出的两个子集和对编号为m=1的模型进行鲁棒训练。Step 1.10: Use SSL technology and contrastive learning technology to divide the two subsets based on Step 1.9. and Perform robust training on the model numbered m=1.
Step 1.11、对于编号为m=2的模型,在每个epoch的开始,首先计算所有训练样本在两个模型上的Jensen–Shannon divergence,并利用小损失准则,对每个类别包含的样本按照其对应的散度值进行排序,并挑选相同数量的样本作为含干净标签的样本,将其索引标记为True值,此外,对于被包含在在基准集合Dc中样本,其索引也被标记为True,所有剩余样本的索引被标记为False值;然后,初始化集合和/>为空,将被标记为True的样本放入干净样本集合/>剩余False样本移除标签后放入/> Step 1.11. For the model with number m=2, at the beginning of each epoch, first calculate the Jensen–Shannon divergence of all training samples on the two models, and use the small loss criterion to sort the samples contained in each category according to their corresponding divergence values, and select the same number of samples as samples with clean labels, and mark their indexes as True values. In addition, for the samples included in the benchmark set D c , their indexes are also marked as True, and the indexes of all remaining samples are marked as False values; then, initialize the set and/> Empty, the samples marked as True are put into the clean sample set/> The remaining False samples are removed from the label and put into />
Step 1.12、使用SSL技术和对比学习技术,基于Step1.11划分出的两个子集和对编号为m=2的模型进行鲁棒训练。Step 1.12: Use SSL technology and contrastive learning technology to divide the two subsets based on Step 1.11. and Robust training is performed on the model numbered m=2.
Step 1.13如果没有达到预设的训练总轮次Ttot,则跳转到step 1.9继续训练。Step 1.13 If the preset total number of training rounds T tot is not reached, jump to step 1.9 to continue training.
实施例2:Embodiment 2:
在实施例1的基础之上,参考附图1示出了本发明的整体学习流程。参考附图2示出了本发明提出的收集基准集Dc过程中的均衡选择的具体流程。Based on Example 1, the overall learning process of the present invention is shown with reference to FIG1. The specific process of the balanced selection in the process of collecting the benchmark set D c proposed by the present invention is shown with reference to FIG2.
设原始数据集为其中n是数据集样本个数,/>是图像xi的观测标签,有一定概率不等于图像的真实标签,c表示数据集包含的类别数。特征提取器和分类器分别表示为h(·;φm)和f(·;θm),其中m∈{1,2}是模型编号,而φm和θm表示编号为m的模型的参数。对于输入图像xi,编号为m的模型的输出预测值可以为p(xi)=f(h(xi;φm);θm)。Assume the original data set is Where n is the number of samples in the data set,/> is the observed label of image x i , which has a certain probability of not being equal to the true label of the image, and c represents the number of categories contained in the dataset. The feature extractor and classifier are denoted as h(·; φ m ) and f(·; θ m ), respectively, where m∈{1,2} is the model number, and φ m and θ m represent the parameters of the model numbered m. For the input image x i , the output prediction value of the model numbered m can be p(x i )=f(h(x i ;φ m );θ m ).
Step 1.1、在利用公式(1)的联合损失函数基于对两个结构完全一致但是初始化参数不一致的DNN模型(模型编号为m={1,2})进行预训练,总epoch数量为Tw;Step 1.1, using the joint loss function of formula (1) based on Pre-train two DNN models (model number is m={1,2}) with the same structure but different initialization parameters, and the total number of epochs is Tw ;
随后,按照公式(2)计算初始噪声数据集中所有样本在编号为m=1的DNN模型下的JSD损失;Then, the JSD loss of all samples in the initial noise data set under the DNN model numbered m=1 is calculated according to formula (2);
此处的pm(xi)=h(f(xi;θm);φm)是编号为m的DNN模型基于图像xi的输出值。而KL(·)表示Kullback-Leibler函数,如公示3所示。Here, p m ( xi ) = h(f( xi ; θm ); φm ) is the output value of the DNN model numbered m based on the image xi . KL(·) represents the Kullback-Leibler function, as shown in Formula 3.
我们基于计算的JSD值,对每个类别的样本按照降序方式进行排序,例如,对第j类包含的样本(j∈{1,2,…,c}),排序后的序列表示为其中sort表示排序函数。基于小损失准则,我们选取前R个样本,将其索引标记为True,其中而dts由公示(4)计算可得:Based on the calculated JSD value, we sort the samples of each category in descending order. For example, for the samples contained in the jth category (j∈{1,2,…,c}), the sorted sequence is expressed as Where sort represents the sorting function. Based on the small loss criterion, we select the first R samples and mark their indexes as True, where And d ts can be calculated from formula (4):
此处的表示所有样本的JSD损失的均值,而/>表示所有样本的JSD的最小值,τ和dμ是预定义的两个超参数。假设当前epoch序号为t,此时可以将所有未被标记为True的样本标记为False,并将所有样本的标记为按照其索引顺序存入历史序列/>中,其中/>为数据集/>中所有样本在第t个epoch时,基于编号为m的模型的标记结果。Here represents the mean of the JSD loss of all samples, and /> represents the minimum value of JSD of all samples, τ and d μ are two predefined hyperparameters. Assuming that the current epoch number is t, all samples that are not marked as True can be marked as False, and the marks of all samples are stored in the history sequence according to their index order/> Among them/> For the dataset/> The labeling results of all samples in the tth epoch based on the model numbered m.
Step 1.2、对于编号为m=1的模型,在第t个epoch的开始,首先按照公式(2)计算所有训练样本在两个模型上的JSD损失,同样利用小损失准则,和step1.1过程一样,对每个类别包含的样本按照其对应的散度值进行排序,并按照公式(5)挑选R个样本作为含干净标签的样本,并标记为True,剩余样本标记为False。随后将标记结果存入在历史序列S1中。最后,初始化集合和/>为空,如公式(6)所示,选择序列S1中连续K个epoch均被标记为True的样本放入干净样本集合/>剩余样本移除标签后放入/> Step 1.2: For the model with number m=1, at the beginning of the tth epoch, first calculate the JSD loss of all training samples on the two models according to formula (2). Also use the small loss criterion, as in step 1.1, sort the samples contained in each category according to their corresponding divergence values, and select R samples as samples with clean labels according to formula (5), and mark them as True, and mark the remaining samples as False. Then store the marking results in the history sequence S1 . Finally, initialize the set and/> is empty, as shown in formula (6), select samples in sequence S1 that are marked as True for K consecutive epochs and put them into the clean sample set/> Remove the labels from the remaining samples and place them in />
其中t表示第t个epoch。Where t represents the tth epoch.
Step 1.3、使用SSL技术和对比学习技术,基于Step1.2划分出的两个子集和对编号为m=1的模型进行鲁棒训练。首先根据公式(7)计算SSL损失Lssl;Step 1.3: Use SSL technology and contrastive learning technology to divide the two subsets based on Step 1.2. and Perform robust training on the model numbered m=1. First, calculate the SSL loss Lssl according to formula (7);
其中,Ll是基于干净标签子集的分类损失,Lul是基于无标签子集/>的分类损失,Lreg是正则损失,其可以促使分类器的预测值不偏向于特定类别。λu是预设的权重系数。此外,本发明引入了SimCLR对比学习方法,对于每个模型m={1,2},其均包含一个映射头/>其中/>是映射头的参数。/>可将特征提取器f(·;θm)提取的高维特征f(xi)映射到低维空间,从而便于进一步区分。基于映射后的低维特征,对于/>中的每个样本,需要按照公式(8)计算对比损失以提高模型的特征提取能力和聚类能力。Among them, L l is based on the clean label subset The classification loss of LuL is based on the unlabeled subset/> The classification loss, L reg is the regularization loss, which can make the prediction value of the classifier not biased towards a specific category. λ u is a preset weight coefficient. In addition, the present invention introduces the SimCLR contrastive learning method, which includes a mapping head for each model m = {1, 2}/> Where/> It is the parameter of the mapping header. /> The high-dimensional features f( xi ) extracted by the feature extractor f(·; θm ) can be mapped to a low-dimensional space, so as to facilitate further distinction. Based on the mapped low-dimensional features, for/> For each sample in , the contrast loss needs to be calculated according to formula (8) to improve the feature extraction and clustering capabilities of the model.
其中是映射后的低维特征;而x2i和x2i-1是样本xi的两次强数据增强变换后的结果。最终,编号为m的模型的总体损失如公式(9)所示:in is the low-dimensional feature after mapping; and x 2i and x 2i-1 are the results of two strong data enhancement transformations of sample xi . Finally, the overall loss of the model numbered m is shown in formula (9):
Step 1.4、对于编号为m=2的模型,在第t个epoch的开始,首先按照公式(2)计算所有训练样本在两个模型上的JSD损失,同样利用小损失准则,和step1.1过程一样,对每个类别包含的样本按照其对应的散度值进行排序,并按照公式(5)挑选R个样本作为含干净标签的样本,并标记为True,剩余样本标记为False。随后将标记结果存入在历史序列S2中。最后,初始化集合和/>为空,根据公式(6)将历史序列S2中连续K个epoch均被标记为True的样本放入干净样本集合/>剩余样本移除标签后放入/> Step 1.4: For the model with number m=2, at the beginning of the tth epoch, first calculate the JSD loss of all training samples on the two models according to formula (2). Also use the small loss criterion, as in step 1.1, sort the samples contained in each category according to their corresponding divergence values, and select R samples as samples with clean labels according to formula (5), and mark them as True, and mark the remaining samples as False. Then store the marking results in the history sequence S2 . Finally, initialize the set and/> is empty, and according to formula (6), the samples that have been marked as True for K consecutive epochs in the historical sequence S 2 are put into the clean sample set/> Remove the labels from the remaining samples and place them in />
Step 1.5、使用SSL技术和对比学习技术,基于Step1.4划分出的两个子集和对编号为m=2的模型进行鲁棒训练,损失函数如公式(9)所示。Step 1.5: Use SSL technology and contrastive learning technology to divide the two subsets based on Step 1.4. and Robust training is performed on the model numbered m=2, and the loss function is shown in formula (9).
Step 1.6如果没有达到预设的训练总轮次Ttot,则跳转到step 1.2继续进行训练。Step 1.6: If the preset total number of training rounds T tot is not reached, jump to step 1.2 to continue training.
Step 1.7步骤1的训练全部结束后,如公式(10)所示,在历史序列S1中最后连续个结果中,挑选包含被标记为True的样本数目最多的那组序列,将该序列中的被标记为True的样本放入基准集合Dc。Step 1.7 After all the training in step 1 is completed, as shown in formula (10), the last continuous Among the results, the sequence with the largest number of samples marked as True is selected, and the samples marked as True in the sequence are put into the reference set D c .
Step 1.8、重新初始化模型编号为m={1,2}的DNN模型,同样使用公式(1)基于给定的含噪声标签数据集进行共Tw轮次(epochs)的温和(warm-up)训练。Step 1.8, re-initialize the DNN model with model number m = {1,2}, and use formula (1) based on the given noisy label dataset Perform warm-up training for a total of T w epochs.
Step 1.9、对于编号为m=1的模型,在第t个epoch的开始,首先按照公式(2)计算所有训练样本在两个模型上的JSD损失,同样利用小损失准则,和step1.1过程一样,对每个类别包含的样本按照其对应的散度值进行排序,并按照公式(5)挑选R个样本作为含干净标签的样本,并标记为True,此外,对于被包含在在基准集合Dc中样本,其索引也被标记为True,所有剩余样本的索引被标记为False值。Step 1.9: For the model with number m=1, at the beginning of the tth epoch, first calculate the JSD loss of all training samples on the two models according to formula (2). Also using the small loss criterion, as in step 1.1, sort the samples contained in each category according to their corresponding divergence values, and select R samples as samples with clean labels according to formula (5) and mark them as True. In addition, for the samples included in the benchmark set D c , their indexes are also marked as True, and the indexes of all remaining samples are marked as False values.
最后,初始化集合和/>为空,根据公式(11)将原始数据集中被标记为True的样本放入干净样本集合/>剩余样本移除标签后放入/> Finally, initialize the collection and/> is empty, and according to formula (11), the samples marked as True in the original data set are put into the clean sample set/> Remove the labels from the remaining samples and place them in />
Step 1.10、使用SSL技术和对比学习技术,基于Step1.9划分出的两个子集和对编号为m=1的模型进行鲁棒训练,损失函数如公式(9)所示。Step 1.10: Use SSL technology and contrastive learning technology to divide the two subsets based on Step 1.9. and Robust training is performed on the model numbered m=1, and the loss function is shown in formula (9).
Step 1.11、对于编号为m=2的模型,在第t个epoch的开始,首先按照公式(2)计算所有训练样本在两个模型上的JSD损失,同样利用小损失准则,和step1.1过程一样,对每个类别包含的样本按照其对应的散度值进行排序,并按照公式(5)挑选R个样本作为含干净标签的样本,并标记为True,此外,对于被包含在在基准集合Dc中样本,其索引也被标记为True,所有剩余样本的索引被标记为False值。最后,初始化集合和/>为空,根据公式(11)将原始数据集中被标记为True的样本放入干净样本集合/>剩余样本移除标签后放入/> Step 1.11, for the model numbered m = 2, at the beginning of the tth epoch, first calculate the JSD loss of all training samples on the two models according to formula (2), and also use the small loss criterion, as in step 1.1, sort the samples contained in each category according to their corresponding divergence values, and select R samples as samples with clean labels according to formula (5) and mark them as True. In addition, for the samples included in the benchmark set D c , their indexes are also marked as True, and the indexes of all remaining samples are marked as False values. Finally, initialize the set and/> is empty, and according to formula (11), the samples marked as True in the original data set are put into the clean sample set/> Remove the labels from the remaining samples and place them in />
Step 1.12、使用SSL技术和对比学习技术,基于Step1.11划分出的两个子集和对编号为m=2的模型进行鲁棒训练,损失函数如公式(9)所示。Step 1.12: Use SSL technology and contrastive learning technology to divide the two subsets based on Step 1.11. and Robust training is performed on the model numbered m=2, and the loss function is shown in formula (9).
Step 1.13如果没有达到预设的训练总轮次Ttot,则跳转到step 1.9继续训练。Step 1.13 If the preset total number of training rounds T tot is not reached, jump to step 1.9 to continue training.
实施例3:Embodiment 3:
本实施例公开一种含噪声标签图像学习系统,该系统包括预训练模块、鲁棒训练模块、挑选模块、重复执行模块。预训练模块利用联合损失函数,基于原始数据集样本,对编号为m={1,2}的DNN模型进行若干轮次的预训练;在历史序列Sm中按照时间顺序,将含干净标签样本的索引标记为True,其他索引标记为False,将标记为True的样本放入子集将剩余样本移除标签后放入子集/>鲁棒训练模块基于所述子集/>子集/>对编号为m={1,2}的DNN模型进行鲁棒训练。挑选模块用于在全部轮次的预训练结束后,在历史序列Sm中最后连续/>个结果中,挑选包含被标记为True的样本数目最多的那组序列,将该序列中的被标记为True的样本放入基准集合Dc。重复执行模块用于重新初始化编号为m={1,2}的DNN模型,反馈至预训练模块、鲁棒训练模块、挑选模块,直到达到预设的训练总轮次。This embodiment discloses a noisy label image learning system, which includes a pre-training module, a robust training module, a selection module, and a repeated execution module. The pre-training module uses a joint loss function to perform several rounds of pre-training on a DNN model numbered m={1,2} based on the original data set samples; in the historical sequence Sm , the index of the sample containing the clean label is marked as True, and the other indexes are marked as False in chronological order, and the samples marked as True are put into the subset Remove labels from the remaining samples and put them into the subset/> The robust training module is based on the subset Subset/> Perform robust training on the DNN model numbered m = {1,2}. The selection module is used to select the last consecutive in the historical sequence S m after all rounds of pre-training are completed. Among the results, the sequence with the largest number of samples marked as True is selected, and the samples marked as True in the sequence are put into the benchmark set D c . The repeated execution module is used to reinitialize the DNN model numbered m = {1, 2}, and feed back to the pre-training module, the robust training module, and the selection module until the preset total number of training rounds is reached.
利用该含噪声标签图像学习系统,可以实现如上述实施例1所公开的基于均衡选择及对比学习的含噪声标签图像学习方法。By using the noisy labeled image learning system, the noisy labeled image learning method based on balanced selection and contrast learning disclosed in the above-mentioned embodiment 1 can be implemented.
实施例4:Embodiment 4:
本实施例提出一种计算机可读存储介质,存储介质中存储有至少一可执行指令,可执行指令在电子设备上运行时,使得电子设备执行如上述实施例所述的基于均衡选择及对比学习的含噪声标签图像学习方法的操作。本公开中的计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑磁盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。计算机可读存储介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了可读程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。可读信号介质还可以是可读存储介质以外的任何可读介质,该可读介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。This embodiment proposes a computer-readable storage medium, in which at least one executable instruction is stored. When the executable instruction is executed on an electronic device, the electronic device performs the operation of the noisy label image learning method based on balanced selection and contrast learning as described in the above embodiment. More specific examples of computer-readable storage media in the present disclosure may include, but are not limited to: an electrical connection with one or more wires, a portable computer disk, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disk read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the above. The computer-readable storage medium may include a data signal propagated in a baseband or as part of a carrier wave, in which a readable program code is carried. Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the above. The readable signal medium may also be any readable medium other than a readable storage medium, which may send, propagate, or transmit a program for use by or in conjunction with an instruction execution system, device, or device.
基于上述实施例所公开的技术方案,本发明提出的基于均衡选择及对比学习的含噪声标签图像学习方法,在合成噪声标签数据集CIFAR-10上,使用相同结构的深度模型作为骨干网络(backbone),增益效果如下:Based on the technical solutions disclosed in the above embodiments, the noisy label image learning method based on balanced selection and contrastive learning proposed in the present invention uses a deep model with the same structure as the backbone network on the synthetic noisy label dataset CIFAR-10, and the gain effect is as follows:
1)90%对称噪声标签场景下,当前最先进的方法DivideMix及LongReMix分别取得93.2%和93.8%的测试精度,而本发明取得了94.0%的精度提升,当对称噪声比率增加到92%时,DivideMix和LongReMix方法测试精度分别为57.6%,63.9%,而本发明测试精度为90.9%;1) In the 90% symmetric noise label scenario, the current most advanced methods DivideMix and LongReMix achieved 93.2% and 93.8% test accuracy respectively, while the present invention achieved 94.0% accuracy improvement. When the symmetric noise ratio increased to 92%, the test accuracy of DivideMix and LongReMix methods were 57.6% and 63.9% respectively, while the test accuracy of the present invention was 90.9%;
2)在49%非对称噪声标签场景下,同类型方法DivideMix和LongReMix精度为83.7%和87.7%,而本发明精度为92.0%,提升效果较为显著;2) In the 49% asymmetric noise label scenario, the accuracy of the similar methods DivideMix and LongReMix is 83.7% and 87.7%, while the accuracy of the present invention is 92.0%, which is a significant improvement;
3)当处于残差(RN)语义噪声标签场景时,DivideMix和LongReMix精度分别为84.57%和85.13%,而本发明精度为92.04%,提升显著。3) When in the residual (RN) semantic noise label scenario, the accuracies of DivideMix and LongReMix are 84.57% and 85.13% respectively, while the accuracy of the present invention is 92.04%, which is a significant improvement.
最后,面对真实噪声标签数据集Animal-10N,竞品方法DivideMix和LongReMix精度分别为84.5%和87.2%,本发明精度为88.84%,提升1.2%。Finally, facing the real noise label dataset Animal-10N, the accuracy of the competing methods DivideMix and LongReMix are 84.5% and 87.2% respectively, and the accuracy of the present invention is 88.84%, an improvement of 1.2%.
如上所述,尽管参照特定的优选实施例已经表示和表述了本发明,但其不得解释为对本发明自身的限制。在不脱离所附权利要求定义的本发明的精神和范围前提下,可对其在形式上和细节上做出各种变化。As described above, although the present invention has been shown and described with reference to specific preferred embodiments, it should not be construed as limiting the present invention itself. Various changes may be made to it in form and detail without departing from the spirit and scope of the present invention as defined in the appended claims.
Claims (10)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202410281796.6A CN118072101B (en) | 2024-03-13 | 2024-03-13 | Noisy label image learning method and system based on balanced selection and contrastive learning |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202410281796.6A CN118072101B (en) | 2024-03-13 | 2024-03-13 | Noisy label image learning method and system based on balanced selection and contrastive learning |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN118072101A true CN118072101A (en) | 2024-05-24 |
| CN118072101B CN118072101B (en) | 2024-11-22 |
Family
ID=91101849
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202410281796.6A Active CN118072101B (en) | 2024-03-13 | 2024-03-13 | Noisy label image learning method and system based on balanced selection and contrastive learning |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN118072101B (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN119091184A (en) * | 2024-07-29 | 2024-12-06 | 江苏开放大学(江苏城市职业学院) | Noisy intestinal ultrasound image classification method and system based on pseudo-label relaxed contrast loss |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20210374553A1 (en) * | 2020-06-02 | 2021-12-02 | Salesforce.Com, Inc. | Systems and methods for noise-robust contrastive learning |
| CN115331088A (en) * | 2022-10-13 | 2022-11-11 | 南京航空航天大学 | Robust learning method based on class labels with noise and imbalance |
| CN115408525A (en) * | 2022-09-29 | 2022-11-29 | 中电科新型智慧城市研究院有限公司 | Petition text classification method, device, equipment and medium based on multi-level label |
| CN116757261A (en) * | 2023-08-16 | 2023-09-15 | 南京航空航天大学 | Robust learning method based on labels with closed set noise and open set noise |
| CN117173494A (en) * | 2023-11-03 | 2023-12-05 | 南京理工智造科技有限公司 | Noise-containing label image recognition method and system based on class balance sample selection |
| CN117197474A (en) * | 2023-09-28 | 2023-12-08 | 江苏开放大学(江苏城市职业学院) | Noise tag learning method based on class equalization and cross combination strategy |
| CN117421657A (en) * | 2023-10-27 | 2024-01-19 | 江苏开放大学(江苏城市职业学院) | Sampling and learning method and system for noisy labels based on oversampling strategy |
-
2024
- 2024-03-13 CN CN202410281796.6A patent/CN118072101B/en active Active
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20210374553A1 (en) * | 2020-06-02 | 2021-12-02 | Salesforce.Com, Inc. | Systems and methods for noise-robust contrastive learning |
| CN115408525A (en) * | 2022-09-29 | 2022-11-29 | 中电科新型智慧城市研究院有限公司 | Petition text classification method, device, equipment and medium based on multi-level label |
| CN115331088A (en) * | 2022-10-13 | 2022-11-11 | 南京航空航天大学 | Robust learning method based on class labels with noise and imbalance |
| CN116757261A (en) * | 2023-08-16 | 2023-09-15 | 南京航空航天大学 | Robust learning method based on labels with closed set noise and open set noise |
| CN117197474A (en) * | 2023-09-28 | 2023-12-08 | 江苏开放大学(江苏城市职业学院) | Noise tag learning method based on class equalization and cross combination strategy |
| CN117421657A (en) * | 2023-10-27 | 2024-01-19 | 江苏开放大学(江苏城市职业学院) | Sampling and learning method and system for noisy labels based on oversampling strategy |
| CN117173494A (en) * | 2023-11-03 | 2023-12-05 | 南京理工智造科技有限公司 | Noise-containing label image recognition method and system based on class balance sample selection |
Non-Patent Citations (1)
| Title |
|---|
| 翟婷婷等: "基于Hubness的类别均衡的时间序列实例选择算法", 计算机应用, vol. 32, no. 11, 1 November 2012 (2012-11-01), pages 3034 - 3037 * |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN119091184A (en) * | 2024-07-29 | 2024-12-06 | 江苏开放大学(江苏城市职业学院) | Noisy intestinal ultrasound image classification method and system based on pseudo-label relaxed contrast loss |
Also Published As
| Publication number | Publication date |
|---|---|
| CN118072101B (en) | 2024-11-22 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11182602B2 (en) | Method and system for person re-identification | |
| CN114492581B (en) | A method based on transfer learning and attention mechanism meta-learning applied to small sample image classification | |
| CN110188827A (en) | A kind of scene recognition method based on convolutional neural networks and recurrence autocoder model | |
| CN110929848B (en) | Training and tracking method based on multi-challenge perception learning model | |
| CN111402257A (en) | An automatic segmentation method of medical images based on multi-task cooperative cross-domain transfer | |
| CN115908947B (en) | A multimodal periodontal image recognition method and system based on inductive learning | |
| CN111191033A (en) | Open set classification method based on classification utility | |
| CN112465700B (en) | A device and method for image stitching and positioning based on depth clustering | |
| CN116543269B (en) | Cross-domain small-sample fine-grained image recognition method and its model based on self-supervision | |
| CN111126361A (en) | SAR target identification method based on semi-supervised learning and feature constraint | |
| CN110853703A (en) | Semi-supervised learning prediction method for protein secondary structure | |
| CN113870863B (en) | Voiceprint recognition method and device, storage medium and electronic equipment | |
| Xi et al. | Robust semi-supervised learning for self-learning open-world classes | |
| CN118072101A (en) | Noisy label image learning method and system based on balanced selection and contrastive learning | |
| CN113705215A (en) | Meta-learning-based large-scale multi-label text classification method | |
| CN113822377A (en) | Fake face detection method based on contrast self-learning | |
| CN116232699A (en) | Training method of fine-grained network intrusion detection model and network intrusion detection method | |
| CN114743133A (en) | A lightweight small sample video classification and recognition method and system | |
| Sajol et al. | A convnext v2 approach to document image analysis: Enhancing high-accuracy classification | |
| CN119380338A (en) | Small sample whole slice pathology image classification method, system and computer readable storage medium based on multimodal large model | |
| CN118799645A (en) | A deep fusion multi-domain few-sample classification method based on knowledge distillation | |
| CN118799948A (en) | Emotion recognition smart contract construction method based on cross fusion and confidence assessment | |
| Shah et al. | Deriving compact feature representations via annealed contraction | |
| Zhang et al. | X-transfer: A transfer learning-based framework for gan-generated fake image detection | |
| CN114863280A (en) | Small sample pest image identification method based on causal intervention technology |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |