CN110223712A - A kind of music emotion recognition method based on two-way convolution loop sparse network - Google Patents
A kind of music emotion recognition method based on two-way convolution loop sparse network Download PDFInfo
- Publication number
- CN110223712A CN110223712A CN201910485792.9A CN201910485792A CN110223712A CN 110223712 A CN110223712 A CN 110223712A CN 201910485792 A CN201910485792 A CN 201910485792A CN 110223712 A CN110223712 A CN 110223712A
- Authority
- CN
- China
- Prior art keywords
- time
- emotion recognition
- convolution
- bcrfms
- audio
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 45
- 230000008909 emotion recognition Effects 0.000 title claims abstract description 40
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 21
- 125000004122 cyclic group Chemical group 0.000 claims abstract description 14
- 230000002996 emotional effect Effects 0.000 claims abstract description 12
- 230000005236 sound signal Effects 0.000 claims abstract description 11
- 238000013528 artificial neural network Methods 0.000 claims abstract description 5
- 238000012549 training Methods 0.000 claims abstract description 5
- 230000002457 bidirectional effect Effects 0.000 claims abstract description 4
- 230000002123 temporal effect Effects 0.000 claims abstract description 3
- 210000002569 neuron Anatomy 0.000 claims description 41
- 230000006870 function Effects 0.000 claims description 30
- 238000010586 diagram Methods 0.000 claims description 15
- 238000012545 processing Methods 0.000 claims description 12
- 230000002441 reversible effect Effects 0.000 claims description 11
- 239000013598 vector Substances 0.000 claims description 11
- 238000006243 chemical reaction Methods 0.000 claims description 10
- 230000008569 process Effects 0.000 claims description 6
- 230000009467 reduction Effects 0.000 claims description 6
- 230000014759 maintenance of location Effects 0.000 claims description 5
- 230000002087 whitening effect Effects 0.000 claims description 5
- 239000011159 matrix material Substances 0.000 claims description 4
- 230000003044 adaptive effect Effects 0.000 claims description 3
- 239000012634 fragment Substances 0.000 claims description 3
- 230000007774 longterm Effects 0.000 claims description 3
- 230000000306 recurrent effect Effects 0.000 claims description 3
- 238000004364 calculation method Methods 0.000 claims description 2
- 230000008859 change Effects 0.000 claims description 2
- 230000004927 fusion Effects 0.000 claims description 2
- 230000015654 memory Effects 0.000 claims description 2
- 238000011176 pooling Methods 0.000 claims description 2
- 230000006403 short-term memory Effects 0.000 claims description 2
- 230000006872 improvement Effects 0.000 description 9
- 230000008451 emotion Effects 0.000 description 6
- 230000037007 arousal Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 239000002360 explosive Substances 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 230000008450 motivation Effects 0.000 description 1
- 238000007500 overflow downdraw method Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0008—Associated control or indicating means
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/27—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique
- G10L25/30—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique using neural networks
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/63—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2210/00—Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
- G10H2210/031—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Acoustics & Sound (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Psychiatry (AREA)
- Hospice & Palliative Care (AREA)
- General Health & Medical Sciences (AREA)
- Child & Adolescent Psychology (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
本发明公开了一种基于双向卷积循环稀疏网络的音乐情感识别方法。本发明结合卷积神经网络和循环神经网络自适应地从原始音频信号的二维时频表示(即时频图)中学习内含时序信息的情感显著性特征。进一步地,本发明提出采用加权混合二进制表示法,通过将回归预测问题转化为多个二分类问题的加权联合来降低数值型真实数据的计算复杂度。实验结果表明,双向卷积循环稀疏网络提取的内含时序信息的情感显著性特征与MediaEval 2015中的最优特征相比表现出更优的预测性能;提出的模型与目前普遍的音乐情感识别网络结构和最优方法相比训练时间减少且预测精度提高。因此,本发明方法有效解决了音乐情感识别的精度和效率的问题,而且优于现有的识别方法。
The invention discloses a music emotion recognition method based on a two-way convolutional cyclic sparse network. The present invention combines the convolutional neural network and the cyclic neural network to adaptively learn the emotional salience features containing time series information from the two-dimensional time-frequency representation (i.e. frequency map) of the original audio signal. Furthermore, the present invention proposes to use a weighted mixed binary representation to reduce the computational complexity of numerical real data by converting the regression prediction problem into a weighted combination of multiple binary classification problems. The experimental results show that the emotional saliency features containing temporal information extracted by the bidirectional convolutional sparse network show better predictive performance than the optimal features in MediaEval 2015; the proposed model is comparable to the current common music emotion recognition network Compared with the optimal method, the training time is reduced and the prediction accuracy is improved. Therefore, the method of the present invention effectively solves the problems of the accuracy and efficiency of music emotion recognition, and is superior to existing recognition methods.
Description
技术领域technical field
本发明属于机器学习与情感计算领域,具体涉及一种基于双向卷积循环稀疏网络的音乐情感识别方法。The invention belongs to the field of machine learning and emotional computing, and in particular relates to a music emotion recognition method based on a two-way convolutional loop sparse network.
背景技术Background technique
随着多媒体技术的发展,来自不同媒体的数字音乐数量的爆炸式增长使得对快速有效的音乐查询和检索方式的研究得到越来越多的关注。由于音乐可以传递情感相关的信息并且基于情感的音乐信息检索方式具有较高的普适性和用户满意度,通过识别音乐音频信号的情感来进行音乐信息检索已经成为了一个重要的研究趋势,其核心难点是如何进一步提高音乐情感识别的精度和效率。With the development of multimedia technology, the explosive growth of the number of digital music from different media has drawn more and more attention to the research on fast and effective music query and retrieval methods. Since music can convey emotion-related information and emotion-based music information retrieval has high universality and user satisfaction, music information retrieval by identifying the emotion of music audio signals has become an important research trend. The core difficulty is how to further improve the accuracy and efficiency of music emotion recognition.
音乐情感识别的目标是通过提取和分析音乐特征如节奏、音色和强度等,学习其感知情感状态。大量基于卷积神经网络(CNN)和循环神经网络(RNN)进行音乐情感的识别研究表现出一定的优越性。CNN可以自适应地从原始音频数据中学习高级不变特征的特性来消除特征提取过程对人类主观性或经验的依赖,RNN可以解决音乐信息的时序依赖问题。本发明采用一种基于双向卷积循环稀疏网络的音乐情感识别方法,结合了CNN自适应学习高级不变特征的特性与RNN学习特征时序关系的能力,用于激励(Arousal)和效价(Valence)情感值的预测,进而提高音乐情感识别的准确度。The goal of music emotion recognition is to learn its perceived emotional state by extracting and analyzing musical features such as tempo, timbre, and intensity. A large number of studies on music emotion recognition based on convolutional neural network (CNN) and recurrent neural network (RNN) have shown certain advantages. CNN can adaptively learn the characteristics of high-level invariant features from raw audio data to eliminate the dependence of the feature extraction process on human subjectivity or experience, and RNN can solve the problem of timing dependence of music information. The present invention adopts a music emotion recognition method based on two-way convolutional cyclic sparse network, which combines the characteristics of CNN self-adaptive learning of advanced invariant features and the ability of RNN to learn the temporal relationship of features, and is used for motivation (Arousal) and valence (Valence ) emotion value prediction, and then improve the accuracy of music emotion recognition.
发明内容Contents of the invention
本发明的目的是提高音乐情感识别的精度和效率,提供了一种基于双向卷积循环稀疏网络的音乐情感识别方法,该方法结合了CNN和RNN来学习时频图中内含时序信息的情感显著性特征,同时采用一种加权混合二进制表示法将回归问题转为多个二分类问题,减少了模型的训练时间并提高了预测精度。The purpose of the present invention is to improve the accuracy and efficiency of music emotion recognition, providing a music emotion recognition method based on two-way convolution cyclic sparse network, which combines CNN and RNN to learn the emotion of timing information contained in the time-frequency graph Salient features, while using a weighted mixed binary representation to convert the regression problem into multiple binary classification problems, reducing the training time of the model and improving the prediction accuracy.
为了达到上述目的,本发明采用如下的技术方案来实现:In order to achieve the above object, the present invention adopts following technical scheme to realize:
一种基于双向卷积循环稀疏网络的音乐情感识别方法,该方法首先将音频信号转化为时频图;其次采用卷积神经网络和循环神经网络内部融合的方式建立音频时序模型来学习内含时序信息的情感显著性特征,简称SII-ASF,同时结合加权混合二进制表示法将回归问题转化为多个二分类问题降低计算复杂度;最后进行音乐的连续情感识别。A music emotion recognition method based on a two-way convolutional cyclic sparse network. This method first converts the audio signal into a time-frequency graph; secondly, the internal fusion of the convolutional neural network and the cyclic neural network is used to establish an audio timing model to learn the internal timing. The emotional saliency feature of information, referred to as SII-ASF, is combined with weighted mixed binary representation to transform the regression problem into multiple binary classification problems to reduce computational complexity; finally, the continuous emotion recognition of music is carried out.
本发明进一步的改进在于,具体包括以下步骤:A further improvement of the present invention is to specifically include the following steps:
1)音频信号的时频图转化:包括音频文件的时频图转化和时频图的降维处理,具体有以下步骤,1) Time-frequency diagram conversion of audio signals: including time-frequency diagram conversion of audio files and dimensionality reduction processing of time-frequency diagrams, the specific steps are as follows,
1-1)音频文件的时频图转化:将每个时域音频文件分割为固定时长的不重叠片段,针对每个片段,设置固定帧长和步长的滑动窗口将其转化为时频图;1-1) Time-frequency map conversion of audio files: each time-domain audio file is divided into non-overlapping segments of fixed duration, and for each segment, a sliding window with fixed frame length and step size is set to convert it into a time-frequency map ;
1-2)时频图的降维处理:采用PCA白化方法,设置99%的数据差异性保留度对时频图的频域维度进行降维;1-2) Dimension reduction processing of time-frequency diagram: PCA whitening method is adopted, and 99% data difference retention is set to reduce the dimensionality of the frequency domain dimension of time-frequency diagram;
2)建立音频时序模型学习内含时序信息的情感显著性特征:结合CNN自适应学习特征和RNN处理时序数据的能力构建双向卷积循环稀疏网络,简称BCRSN;通过CNN局部互连和权值共享的方式来改变模型输入层与隐藏层之间的连接,使用多个卷积核来获得双向卷积循环特征图组,简称BCRFMs;通过长短时记忆网络(LSTM)模块代替BCRFMs中的每个神经元来考虑BCRFMs之间的长期依赖关系,长短时记忆网络简称LSTM;2) Establish an audio timing model to learn emotional salience features that contain timing information: combine CNN adaptive learning features and RNN's ability to process timing data to build a bidirectional convolutional cyclic sparse network, referred to as BCRSN; through CNN local interconnection and weight sharing The way to change the connection between the input layer and the hidden layer of the model is to use multiple convolution kernels to obtain the two-way convolution cycle feature map group, referred to as BCRFMs; replace each neuron in BCRFMs by the long short-term memory network (LSTM) module To consider the long-term dependencies between BCRFMs, the long-short-term memory network is referred to as LSTM;
3)回归问题转化为二分类问题:包括二进制数值的表示和稀疏处理,具体有以下步骤,3) The regression problem is transformed into a binary classification problem: including the representation and sparse processing of binary values, the specific steps are as follows,
3-1)二进制数值的表示:基于数值型真实数据的表示方法,加权混合二进制表示法,将回归问题转化为多个二分类问题的加权联合以降低模型的计算复杂度;3-1) Representation of binary values: Based on the representation method of numerical real data, the weighted mixed binary representation method converts the regression problem into a weighted combination of multiple binary classification problems to reduce the computational complexity of the model;
3-2)稀疏处理:使用一致性相关系数作为损失函数并向CCC中增加惩罚项作为模型的目标函数来使得BCRFMs尽可能稀疏,获取SII-ASF,其中一致性相关系数简称CCC;3-2) Sparse processing: use the consistency correlation coefficient as the loss function and add a penalty term to CCC as the objective function of the model to make BCRFMs as sparse as possible, and obtain SII-ASF, where the consistency correlation coefficient is referred to as CCC;
4)音乐的连续情感识别:根据多个二分类的结果先对一个片段的音频内容进行情感识别,再对完整音乐文件的多个音频片段进行连续的情感识别。4) Continuous emotion recognition of music: According to the results of multiple binary classifications, the emotion recognition is first performed on the audio content of a segment, and then the continuous emotion recognition is performed on multiple audio segments of the complete music file.
本发明进一步的改进在于,所述步骤1-1)具体操作为:以时长500ms的单位将每个时域音频文件分割为不重叠的片段,对于每一个分割后的音频片段,采用60ms帧长和10ms步长的滑动窗口将其转化为时频图。A further improvement of the present invention is that the specific operation of the step 1-1) is: each time-domain audio file is divided into non-overlapping segments in units of 500 ms in duration, and for each segmented audio segment, a frame length of 60 ms is used and a sliding window with a step size of 10ms to convert it into a time-frequency plot.
本发明进一步的改进在于,所述步骤1-2)具体操作为:以99%的数据差异性保留度进行PCA白化,将时频图频域的维度降低到45维,得到45×45大小的时频图作为BCRSN模型的输入。The further improvement of the present invention is that the specific operation of the step 1-2) is: perform PCA whitening with 99% data difference retention, reduce the dimension of the frequency domain of the time-frequency map to 45 dimensions, and obtain a 45×45 size Time-frequency plots are used as input to the BCRSN model.
本发明进一步的改进在于,所述步骤2)具体操作为:使用64个3×1且步长为2的卷积核对时频图做时域范围内的卷积操作得到BCRFMs;BCRFMs内神经元之间存在按照音频帧的时间顺序的双向循环,某一帧的神经元的输入是其对应卷积结果与前/后一帧的神经元输出的加权和;同时利用LSTM模块修改BCRFMs中的每个神经元,通过该模块的输入、输出和遗忘门限来记忆任意时长片段的某一信息,最后用3×1大小的下采样操作降低特征图尺寸,加强模型的鲁棒性。The further improvement of the present invention is that the specific operation of the step 2) is: use 64 3*1 convolution kernels with a step size of 2 to perform convolution operations on the time-frequency map in the time domain to obtain BCRFMs; neurons in BCRFMs There is a two-way loop according to the time sequence of the audio frame. The input of a neuron in a certain frame is the weighted sum of its corresponding convolution result and the neuron output of the previous/next frame; at the same time, the LSTM module is used to modify each of the BCRFMs. A neuron, through the input, output and forgetting threshold of the module to memorize a certain information of any length segment, and finally use the 3×1 downsampling operation to reduce the size of the feature map and enhance the robustness of the model.
本发明进一步的改进在于,步骤2)中BCRFMs的学习,包含以下步骤:A further improvement of the present invention is that the learning of BCRFMs in step 2) comprises the following steps:
(i)BCRSN模型输入层与正向和反向卷积循环层之间的连接以卷积核为媒介,正向和反向卷积循环层设置与CNN卷积层相同的神经元个数及排列方式,使得模型具有自适应学习不变特征的能力,通过公式(1)计算每个神经元的卷积结果:(i) The connection between the input layer of the BCRSN model and the forward and reverse convolutional loop layers uses the convolution kernel as the medium, and the forward and reverse convolutional loop layers are set with the same number of neurons as the CNN convolutional layer and The arrangement makes the model have the ability to adaptively learn invariant features, and the convolution result of each neuron is calculated by formula (1):
式中,Cnt,k为第k个特征图位置(n,t)处神经元的卷积结果,n=1,2,...,(N-1)/2,t=1,2...,T;为输入层对应位置(n,t)处的二维特征矩阵,Wk为第k个卷积核的权值参数;In the formula, C nt,k is the convolution result of the neuron at the kth feature map position (n,t), n=1,2,...,(N-1)/2, t=1,2 ..., T; is the two-dimensional feature matrix at the corresponding position (n, t) of the input layer, and W k is the weight parameter of the kth convolution kernel;
(ii)BCRFMs内神经元之间存在按照音频帧的时间顺序的双向循环,某一帧的神经元的输入是其对应卷积结果与前/后一帧的神经元输出的加权和;(ii) There is a two-way cycle between neurons in BCRFMs according to the time sequence of audio frames, and the input of a neuron in a certain frame is the weighted sum of its corresponding convolution result and the neuron output of the previous/next frame;
对于正向卷积循环层的特征图,每个神经元的输入用公式(2)表示:For the feature map of the forward convolution loop layer, the input of each neuron is expressed by formula (2):
输出表示为公式(3):The output is expressed as formula (3):
FOnt,k=σ(FInt,k+bnt,k) (3)FO nt,k = σ(FI nt,k +b nt,k ) (3)
对于反向卷积循环层的特征图,每个神经元的输入用公式(4)表示:For the feature map of the reverse convolutional recurrent layer, the input of each neuron is expressed by formula (4):
输出表示为公式(5):The output is expressed as equation (5):
BOnt,k=σ(BInt,k+bnt,k) (5)BO nt,k = σ(BI nt,k +b nt,k ) (5)
式中表示第k个特征图前一帧t-1/t+1的所有神经元的输出结果; 分别表示前向传播和后向传播过程中神经元的连接矩阵,各个音频帧之间共享权值;bnt,k为网络偏置;In the formula Represents the output results of all neurons in the previous frame t-1/t+1 of the kth feature map; Represents the connection matrix of neurons in the process of forward propagation and backward propagation, and the weights are shared between each audio frame; b nt,k is the network bias;
(iii)使用LSTM模块修改BCRFMs中的每个神经元,通过该模块的输入、输出和遗忘门限来记忆任意时长片段的某一信息,在正向和反向卷积循环层与正向和反向池化层之间在频域范围内进行下采样操作,用3×1大小的下采样区域内的最大特征来依次代表该区域的特征,降低特征图大小。(iii) Use the LSTM module to modify each neuron in the BCRFMs, through the input, output and forgetting threshold of the module to remember a certain information of any length of time, in the forward and reverse convolution layer and the forward and reverse The downsampling operation is performed in the frequency domain between the pooling layers, and the largest feature in the downsampling area of 3×1 size is used to represent the features of the area in turn, reducing the size of the feature map.
本发明进一步的改进在于,所述步骤3-1)具体操作为:在BCRSN模型输出层设置L+1个神经元,得到的预测序列用O表示;其中,O1预测真实值的正负,O2~OL+1预测真实值的绝对值大小,其范围在(0,1);每个神经元作为一个二分类器,从而将损失函数计算复杂度降低为O((L+1)×12)=O(L+1),使得模型收敛更快。The further improvement of the present invention is that the specific operation of said step 3-1) is: L+ 1 neurons are set at the output layer of the BCRSN model, and the obtained prediction sequence is represented by O; wherein, O1 predicts the positive or negative of the true value, O 2 ~ OL+1 predicts the absolute value of the real value, and its range is (0,1); each neuron acts as a binary classifier, thereby reducing the computational complexity of the loss function to O((L+1) ×1 2 )=O(L+1), which makes the model converge faster.
本发明进一步的改进在于,步骤3-1)中采用加权混合二进制数值表示方法,包含以下步骤:A further improvement of the present invention is that in step 3-1), a weighted mixed binary value representation method is adopted, comprising the following steps:
(i)新的加权混合二进制表示法将数值型真实数据g转换为混合二进制向量O*来降低计算复杂度,该向量的每一位用公式(6)计算得到:(i) The new weighted mixed binary representation converts the numerical real data g into a mixed binary vector O * to reduce the computational complexity, each bit of the vector Calculated with formula (6):
式中g1=g;由g1值的正负决定,当g1≥0时,g1<0时, where g 1 =g; Determined by the positive or negative value of g 1 , when g 1 ≥ 0, When g 1 <0,
(ii)设置输出层神经元Oi对模型损失函数的贡献权值来控制其收敛方向,提高预测精度,通过下式计算:(ii) Set the contribution weight of the output layer neuron O i to the model loss function to control its convergence direction and improve the prediction accuracy, which is calculated by the following formula:
式中δ(·)表示损失函数的计算公式,λi表示Oi对该片段损失函数的贡献。where δ( ) represents the calculation formula of the loss function, and λi represents the contribution of O i to the segment loss function.
本发明进一步的改进在于,所述步骤3-2)具体操作为:使用CCC作为损失函数并向CCC中增加BCRFMs权值的Lasso惩罚项作为模型的目标函数来使得BCRFMs尽可能稀疏,获取SII-ASF。A further improvement of the present invention is that the step 3-2) is specifically operated as: using CCC as a loss function and adding the Lasso penalty item of the BCRFMs weight to the CCC as the objective function of the model to make the BCRFMs as sparse as possible, and obtain SII- ASF.
本发明进一步的改进在于,步骤3-2)中以CCC作为损失函数以使网络得到更有区分性的训练;具体地,将每首歌分割为固定时长的片段且每个片段的真实数据转化为混合二进制向量O*,损失函数求解包含以下步骤:A further improvement of the present invention is that in step 3-2), CCC is used as a loss function so that the network can be trained more discriminatively; specifically, each song is divided into segments of fixed duration and the real data of each segment is transformed into For mixed binary vectors O * , the loss function solution consists of the following steps:
(i)计算每个片段预测序列O和真实序列O*的CCC,序列样本s的预测序列fs和目标序列之间的CCC定义为:(i) Calculate the CCC of the predicted sequence O and the true sequence O * for each fragment, the predicted sequence f s and the target sequence of the sequence sample s The CCC between is defined as:
式中Ss表示和方误差(SSE),Qs表示协方差,t表示每个标记值的时间索引,Ns表示序列s长度;基于此,以混合二进制向量的位数L+1作为每个片段的序列长度并考虑每一位对模型损失函数的贡献权值,重写公式(7)得到每个片段预测序列O和真实序列O*的CCC:where S s represents the sum square error (SSE), Q s represent the covariance, t represents the time index of each tag value, and N s represents the length of the sequence s; based on this, the number of bits L+1 of the mixed binary vector is used as the sequence length of each segment and the contribution weight of each bit to the model loss function is considered , rewrite formula (7) to get the CCC of each fragment prediction sequence O and real sequence O * :
式中,O*,O分别表示片段真实的和预测的混合二进制向量,λ=(λ1,λ2,...,λL+1)表示O对片段损失函数的贡献参数集合;因此,将回归预测问题的CCC求解转化为多个二分类准确率的加权和,即从而定义:where O * , O represent the real and predicted mixed binary vectors of the segment, respectively, and λ=(λ 1 ,λ 2 ,...,λ L+1 ) represents the set of contribution parameters of O to the segment loss function; therefore, The CCC solution of the regression prediction problem is transformed into the weighted sum of multiple binary classification accuracy rates, namely Thus defining:
(ii)计算每首歌的平均CCC,由其每个片段的CCC和片段数计算得到:(ii) Calculate the average CCC of each song, which is calculated from the CCC and the number of segments of each segment:
式中,Ns表示每首歌曲的长度,即片段数;In the formula, N s represents the length of each song, i.e. the number of segments;
利用Lasso回归将某些神经元的系数置为0来删除重复相关的变量和许多噪声特征,选择出情感显著性更强的SII-ASF;具体地,在损失函数的基础上添加BCRFMs权值的Lasso惩罚项作为最终的目标函数:Use Lasso regression to set the coefficients of some neurons to 0 to delete repeated related variables and many noise features, and select SII-ASF with stronger emotional significance; specifically, in the loss function On the basis of adding the Lasso penalty item of BCRFMs weight as the final objective function:
式中,βF表示BCRFMs的参数集合,类似的,αF和αB是用来控制特征图稀疏度的超参数,α值越大,稀疏度越高;最小化L以删除噪声特征,选择出情感显著性特征,同时提高预测准确度。In the formula, β F represents the parameter set of BCRFMs, akin, α F and α B are hyperparameters used to control the sparsity of feature maps. The larger the value of α, the higher the sparsity; minimize L to delete noise features, select emotionally salient features, and improve prediction accuracy.
本发明具有如下有益的技术效果:The present invention has following beneficial technical effect:
本发明提供的一种基于双向卷积循环稀疏网络的音乐情感识别方法,首先将音频信号转化为时频图,其次采用CNN和RNN内部融合的方式建立音频时序模型来学习SII-ASF,同时结合加权混合二进制表示法将回归问题转化为多个二分类问题降低计算复杂度,最后进行音乐的连续情感识别。与目前普遍的音乐情感识别网络结构和最优方法相比,BCRSN模型可以明显减少训练时间并提高预测精度,提取的SII-ASF特征相比于MediaEval 2015中参赛者提出的最优特征表现出更优的预测性能。A music emotion recognition method based on a two-way convolutional cyclic sparse network provided by the present invention first converts the audio signal into a time-frequency graph, and secondly adopts the internal integration of CNN and RNN to establish an audio timing model to learn SII-ASF, and at the same time combines The weighted mixed binary representation transforms the regression problem into multiple binary classification problems to reduce computational complexity, and finally performs continuous emotion recognition of music. Compared with the current common music emotion recognition network structure and optimal methods, the BCRSN model can significantly reduce the training time and improve the prediction accuracy, and the extracted SII-ASF features are better than the optimal features proposed by the participants in MediaEval 2015. excellent predictive performance.
附图说明Description of drawings
图1为本发明中BCRSN系统流程图;Fig. 1 is the flow chart of BCRSN system among the present invention;
图2为本发明中从数值型真实数据到混合二进制向量的转换过程图;Fig. 2 is the conversion process figure from numerical real data to mixed binary vector among the present invention;
图3为本发明中在DEAM和MTurk音乐情感识别数据集上,BCRSN模型与基于CNN、基于BLSTM以及基于stacked CNN-BLSTM的模型在预测性能和训练时间方面的对比图。Fig. 3 is a comparison chart of prediction performance and training time between the BCRSN model and CNN-based, BLSTM-based and stacked CNN-BLSTM-based models on the DEAM and MTurk music emotion recognition data sets in the present invention.
具体实施方式Detailed ways
下面结合附图对本发明做进一步详细描述。The present invention will be described in further detail below in conjunction with the accompanying drawings.
参照图1,本发明提供的一种基于双向卷积循环稀疏网络的音乐情感识别方法,首先将音频信号转化为时频图;其次采用卷积神经网络(CNN)和循环神经网络(RNN)内部融合的方式建立音频时序模型来学习内含时序信息的情感显著性特征(简称SII-ASF),同时结合加权混合二进制表示法将回归问题转化为多个二分类问题降低计算复杂度;最后进行音乐的连续情感识别,具体包括以下步骤:With reference to Fig. 1, a kind of music emotion recognition method based on two-way convolution cyclic sparse network provided by the present invention first converts audio signal into a time-frequency graph; secondly adopts convolutional neural network (CNN) and cyclic neural network (RNN) internal The fusion method is used to build an audio timing model to learn the emotional saliency feature (SII-ASF) containing timing information, and at the same time combine the weighted mixed binary representation to convert the regression problem into multiple binary classification problems to reduce the computational complexity; finally, the music Continuous emotion recognition, which specifically includes the following steps:
1)音频信号的时频图转化:包括音频文件的时频图转化和对时频图的降维处理,具体有以下步骤,1) Time-frequency diagram conversion of audio signals: including time-frequency diagram conversion of audio files and dimensionality reduction processing of time-frequency diagrams, the specific steps are as follows,
Step1音频文件的时频图转化:将每个时域音频文件分割为固定时长的不重叠片段,针对每个片段,设置固定帧长和步长的滑动窗口将其转化为时频图;Step1 Time-frequency map conversion of audio files: each time-domain audio file is divided into non-overlapping segments of fixed duration, and for each segment, a sliding window with fixed frame length and step size is set to convert it into a time-frequency map;
Step2时频图的降维处理:采用PCA白化方法,设置一定的数据差异性保留度对时频图的频域维度进行降维。Step2 Dimensionality reduction processing of time-frequency diagram: PCA whitening method is used to set a certain degree of data difference retention to reduce the dimensionality of the frequency domain dimension of the time-frequency diagram.
2)建立音频时序模型学习内含时序信息的情感显著性特征:结合CNN自适应学习特征和RNN处理时序数据的能力构建双向卷积循环稀疏网络(简称BCRSN)。参照图1,将输入的二维时频图通过CNN局部互连和权值共享的方式来代替每一帧ti内输入层和正向和反向卷积循环层(Forward/Backward 1c Layer)的层间连接,并且音频帧之间 设置双向循环传递时序信息来学习BCRFMs;同时使用LSTM网络模块代替BCRFMs中的每个神经元,使得BCRFMs内的特征之间具有长期依赖关系。2) Establish an audio timing model to learn emotional salience features that contain timing information: Combine CNN adaptive learning features and RNN's ability to process timing data to build a bidirectional convolutional cyclic sparse network (BCRSN for short). Referring to Figure 1, the input two-dimensional time-frequency image is replaced by CNN local interconnection and weight sharing in the input layer and forward and reverse convolution loop layers (Forward/Backward 1c Layer) in each frame t i between layers, and between audio frames Set up two-way loop transfer timing information to learn BCRFMs; at the same time, use LSTM network module to replace each neuron in BCRFMs, so that there is a long-term dependency between the features in BCRFMs.
3)回归问题转化为二分类问题:包括加权二进制数值的表示和稀疏处理,参照图1和图2,具体有以下步骤,3) The regression problem is converted into a binary classification problem: including the representation and sparse processing of weighted binary values, referring to Figure 1 and Figure 2, the specific steps are as follows,
Step1加权二进制数值的表示:基于表示数值型真实数据的方法,加权混合二进制表示法,将回归问题转化为多个二分类问题的加权联合以降低计算的复杂度;Step1 Representation of weighted binary values: Based on the method of representing numerical real data, the weighted mixed binary representation converts the regression problem into a weighted combination of multiple binary classification problems to reduce the computational complexity;
Step2稀疏处理:使用CCC作为损失函数并向CCC中增加BCRFMs权值的Lasso惩罚项(L1正则化)作为模型的目标函数来使得BCRFMs尽可能稀疏,获取SII-ASF。Step2 sparse processing: use CCC as the loss function and add the Lasso penalty term (L1 regularization) of the weight of BCRFMs to CCC as the objective function of the model to make BCRFMs as sparse as possible and obtain SII-ASF.
4)音乐的连续情感识别:将音频时频图输入BCRSN模型中,根据多个二分类的结果先对单个片段的音频内容进行情感识别,再对完整音乐文件的多个音频片段进行连续的情感识别。4) Continuous emotion recognition of music: Input the audio time-frequency graph into the BCRSN model, first perform emotion recognition on the audio content of a single segment according to the results of multiple binary classifications, and then perform continuous emotion recognition on multiple audio segments of the complete music file identify.
参照图3,在DEAM和MTurk数据集上,本发明中的BCRSN模型与基于CNN、基于BLSTM以及基于stacked CNN-BLSTM的模型相比,Valence和Arousal维度的连续情感预测均取得最优性能。Referring to Figure 3, on the DEAM and MTurk datasets, the BCRSN model in the present invention achieves the best performance in continuous emotion prediction in both Valence and Arousal dimensions compared with CNN-based, BLSTM-based and stacked CNN-BLSTM-based models.
参照表1,与MediaEval 2015的最优算法相比,本发明中的BCRSN模型可以在先验知识最少的情况下针对预测目标从原始音频信号中自适应的学习有效的特征,优于MediaEval2015中的前三个性能最优的方法(BLSTM-RNN、BLSTM-ELM和deep LSTM-RNN)。Referring to Table 1, compared with the optimal algorithm of MediaEval 2015, the BCRSN model in the present invention can learn effective features from the original audio signal for the prediction target with the least prior knowledge, which is better than that of MediaEval2015 Top three best performing methods (BLSTM-RNN, BLSTM-ELM and deep LSTM-RNN).
表1:本发明中以原始音频信号为输入时,BCRSN模型与MediaEval 2015中的前三个性能最优的方法(BLSTM-RNN、BLSTM-ELM和deep LSTM-RNN)的对比。Table 1: Comparison of the BCRSN model with the top three best-performing methods (BLSTM-RNN, BLSTM-ELM and deep LSTM-RNN) in MediaEval 2015 when the original audio signal is used as input in the present invention.
注:N.S.-Not Significant表示该方法的性能与BCRSN模型比无显著性差异,否则表示有显著性差异。Note: N.S.-Not Significant means that the performance of this method has no significant difference compared with the BCRSN model, otherwise it means that there is a significant difference.
参照表2,本发明中BCRSN模型在有Lasso惩罚项和无Lasso惩罚项时得到的SII-ASF和SII-NASF相比于MediaEval 2015中参赛者提出的特征集(JUNLP、PKUAIPL、HKPOLYU、THU-HCSIL和IRIT-SAMOVA),均表现出良好预测性能。Referring to Table 2, the SII-ASF and SII-NASF obtained by the BCRSN model in the present invention when there is a Lasso penalty item and no Lasso penalty item are compared to the feature set (JUNLP, PKUAIPL, HKPOLYU, THU- HCSIL and IRIT-SAMOVA), both showed good predictive performance.
表2:本发明中提取的SII-ASF和SII-NASF特征与MediaEval 2015中参赛者提出的特征(JUNLP、PKUAIPL、HKPOLYU、THU-HCSIL和IRIT-SAMOVA)的性能对比。Table 2: Performance comparison between the SII-ASF and SII-NASF features extracted in the present invention and the features proposed by the competitors in MediaEval 2015 (JUNLP, PKUAIPL, HKPOLYU, THU-HCSIL and IRIT-SAMOVA).
注:N.S.-Not Significant表示该特征的性能与SII-ASF比无显著性差异,否则表示有显著性差异。Note: N.S.-Not Significant means that there is no significant difference between the performance of this feature and the SII-ASF ratio, otherwise it means that there is a significant difference.
Claims (10)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201910485792.9A CN110223712B (en) | 2019-06-05 | 2019-06-05 | A music emotion recognition method based on bidirectional convolutional recurrent sparse network |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201910485792.9A CN110223712B (en) | 2019-06-05 | 2019-06-05 | A music emotion recognition method based on bidirectional convolutional recurrent sparse network |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN110223712A true CN110223712A (en) | 2019-09-10 |
| CN110223712B CN110223712B (en) | 2021-04-20 |
Family
ID=67819412
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201910485792.9A Active CN110223712B (en) | 2019-06-05 | 2019-06-05 | A music emotion recognition method based on bidirectional convolutional recurrent sparse network |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN110223712B (en) |
Cited By (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN110689902A (en) * | 2019-12-11 | 2020-01-14 | 北京影谱科技股份有限公司 | Audio signal time sequence processing method, device and system based on neural network and computer readable storage medium |
| CN111326164A (en) * | 2020-01-21 | 2020-06-23 | 大连海事大学 | Semi-supervised music melody extraction method |
| CN113268628A (en) * | 2021-04-14 | 2021-08-17 | 上海大学 | Music emotion recognition method based on modularized weighted fusion neural network |
| CN115294644A (en) * | 2022-06-24 | 2022-11-04 | 北京昭衍新药研究中心股份有限公司 | Rapid monkey behavior identification method based on 3D convolution parameter reconstruction |
| CN118298799A (en) * | 2024-06-06 | 2024-07-05 | 清华大学 | Low-delay generation audio detection continuous learning method, device, equipment and medium based on sparse sliding window |
Citations (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN105469065A (en) * | 2015-12-07 | 2016-04-06 | 中国科学院自动化研究所 | Recurrent neural network-based discrete emotion recognition method |
| CN106128479A (en) * | 2016-06-30 | 2016-11-16 | 福建星网视易信息系统有限公司 | A kind of performance emotion identification method and device |
| CN106228977A (en) * | 2016-08-02 | 2016-12-14 | 合肥工业大学 | The song emotion identification method of multi-modal fusion based on degree of depth study |
| US9570091B2 (en) * | 2012-12-13 | 2017-02-14 | National Chiao Tung University | Music playing system and music playing method based on speech emotion recognition |
| WO2017122798A1 (en) * | 2016-01-14 | 2017-07-20 | 国立研究開発法人産業技術総合研究所 | Target value estimation system, target value estimation method, and target value estimation program |
| CN107169409A (en) * | 2017-03-31 | 2017-09-15 | 北京奇艺世纪科技有限公司 | A kind of emotion identification method and device |
| CN107506722A (en) * | 2017-08-18 | 2017-12-22 | 中国地质大学(武汉) | One kind is based on depth sparse convolution neutral net face emotion identification method |
| US20180075343A1 (en) * | 2016-09-06 | 2018-03-15 | Google Inc. | Processing sequences using convolutional neural networks |
| CN108717856A (en) * | 2018-06-16 | 2018-10-30 | 台州学院 | A kind of speech-emotion recognition method based on multiple dimensioned depth convolution loop neural network |
| CN108805089A (en) * | 2018-06-14 | 2018-11-13 | 南京云思创智信息科技有限公司 | Based on multi-modal Emotion identification method |
| CN109147826A (en) * | 2018-08-22 | 2019-01-04 | 平安科技(深圳)有限公司 | Music emotion recognition method, device, computer equipment and computer storage medium |
| CN109146066A (en) * | 2018-11-01 | 2019-01-04 | 重庆邮电大学 | A kind of collaborative virtual learning environment natural interactive method based on speech emotion recognition |
| CN109508375A (en) * | 2018-11-19 | 2019-03-22 | 重庆邮电大学 | A kind of social affective classification method based on multi-modal fusion |
| CN109599128A (en) * | 2018-12-24 | 2019-04-09 | 北京达佳互联信息技术有限公司 | Speech-emotion recognition method, device, electronic equipment and readable medium |
-
2019
- 2019-06-05 CN CN201910485792.9A patent/CN110223712B/en active Active
Patent Citations (15)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US9570091B2 (en) * | 2012-12-13 | 2017-02-14 | National Chiao Tung University | Music playing system and music playing method based on speech emotion recognition |
| CN105469065B (en) * | 2015-12-07 | 2019-04-23 | 中国科学院自动化研究所 | A Discrete Emotion Recognition Method Based on Recurrent Neural Network |
| CN105469065A (en) * | 2015-12-07 | 2016-04-06 | 中国科学院自动化研究所 | Recurrent neural network-based discrete emotion recognition method |
| WO2017122798A1 (en) * | 2016-01-14 | 2017-07-20 | 国立研究開発法人産業技術総合研究所 | Target value estimation system, target value estimation method, and target value estimation program |
| CN106128479A (en) * | 2016-06-30 | 2016-11-16 | 福建星网视易信息系统有限公司 | A kind of performance emotion identification method and device |
| CN106228977A (en) * | 2016-08-02 | 2016-12-14 | 合肥工业大学 | The song emotion identification method of multi-modal fusion based on degree of depth study |
| US20180075343A1 (en) * | 2016-09-06 | 2018-03-15 | Google Inc. | Processing sequences using convolutional neural networks |
| CN107169409A (en) * | 2017-03-31 | 2017-09-15 | 北京奇艺世纪科技有限公司 | A kind of emotion identification method and device |
| CN107506722A (en) * | 2017-08-18 | 2017-12-22 | 中国地质大学(武汉) | One kind is based on depth sparse convolution neutral net face emotion identification method |
| CN108805089A (en) * | 2018-06-14 | 2018-11-13 | 南京云思创智信息科技有限公司 | Based on multi-modal Emotion identification method |
| CN108717856A (en) * | 2018-06-16 | 2018-10-30 | 台州学院 | A kind of speech-emotion recognition method based on multiple dimensioned depth convolution loop neural network |
| CN109147826A (en) * | 2018-08-22 | 2019-01-04 | 平安科技(深圳)有限公司 | Music emotion recognition method, device, computer equipment and computer storage medium |
| CN109146066A (en) * | 2018-11-01 | 2019-01-04 | 重庆邮电大学 | A kind of collaborative virtual learning environment natural interactive method based on speech emotion recognition |
| CN109508375A (en) * | 2018-11-19 | 2019-03-22 | 重庆邮电大学 | A kind of social affective classification method based on multi-modal fusion |
| CN109599128A (en) * | 2018-12-24 | 2019-04-09 | 北京达佳互联信息技术有限公司 | Speech-emotion recognition method, device, electronic equipment and readable medium |
Non-Patent Citations (5)
| Title |
|---|
| B SUN: ""LSTM for dynamic emotion and group emotion recognition in the wild"", 《THE 18TH ACM INTERNATIONAL CONFERENCE 》 * |
| M MALIK: ""stacked convolutional recurrent neural networks for music emotion recognition"", 《ARXIV:1706.02292V1》 * |
| YANG XINYU: ""review of data features-based music Emotion Recognition method"", 《MULTIMEDIA SYSTEM》 * |
| 唐霞: ""基于深度学习的音乐情感识别"", 《电脑知识与技术》 * |
| 张昕然: ""跨库语音情感识别若干关键技术研究"", 《中国博士学位论文全文数据库信息科技辑》 * |
Cited By (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN110689902A (en) * | 2019-12-11 | 2020-01-14 | 北京影谱科技股份有限公司 | Audio signal time sequence processing method, device and system based on neural network and computer readable storage medium |
| CN111326164A (en) * | 2020-01-21 | 2020-06-23 | 大连海事大学 | Semi-supervised music melody extraction method |
| CN111326164B (en) * | 2020-01-21 | 2023-03-21 | 大连海事大学 | Semi-supervised music theme extraction method |
| CN113268628A (en) * | 2021-04-14 | 2021-08-17 | 上海大学 | Music emotion recognition method based on modularized weighted fusion neural network |
| CN113268628B (en) * | 2021-04-14 | 2023-05-23 | 上海大学 | Music emotion recognition method based on modularized weighted fusion neural network |
| CN115294644A (en) * | 2022-06-24 | 2022-11-04 | 北京昭衍新药研究中心股份有限公司 | Rapid monkey behavior identification method based on 3D convolution parameter reconstruction |
| CN118298799A (en) * | 2024-06-06 | 2024-07-05 | 清华大学 | Low-delay generation audio detection continuous learning method, device, equipment and medium based on sparse sliding window |
| CN118298799B (en) * | 2024-06-06 | 2024-08-13 | 清华大学 | Low-latency generation audio detection continuous learning method, device, equipment and medium based on sparse sliding window |
Also Published As
| Publication number | Publication date |
|---|---|
| CN110223712B (en) | 2021-04-20 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN109271522B (en) | Review sentiment classification method and system based on deep hybrid model transfer learning | |
| CN110223712A (en) | A kind of music emotion recognition method based on two-way convolution loop sparse network | |
| CN109299342B (en) | Cross-modal retrieval method based on cycle generation type countermeasure network | |
| CN107516129B (en) | Dimension self-adaptive Tucker decomposition-based deep network compression method | |
| CN112507898A (en) | Multi-modal dynamic gesture recognition method based on lightweight 3D residual error network and TCN | |
| CN111105032A (en) | GAN-based chromosome structure anomaly detection method, system and storage medium | |
| CN113688878A (en) | A Small-Sample Image Classification Method Based on Memory Mechanism and Graph Neural Network | |
| CN117407772B (en) | Method and system for classifying training multi-element time sequence data by supervising and comparing learning network model | |
| Wang et al. | Cross-modal retrieval: a systematic review of methods and future directions | |
| CN114926770A (en) | Video motion recognition method, device, equipment and computer readable storage medium | |
| CN113643724B (en) | Kiwi emotion recognition method and system based on time-frequency double-branch characteristics | |
| CN116210009A (en) | Dynamic Quantization for Energy Efficient Deep Learning | |
| CN114866172B (en) | Interference identification method and device based on inverse residual deep neural network | |
| CN114049527B (en) | Self-knowledge distillation method and system based on online collaboration and fusion | |
| CN115687772A (en) | Sequence recommendation method based on sequence dependence enhanced self-attention network | |
| CN116341558A (en) | Multi-modal emotion recognition method and model based on multi-level graph neural network | |
| CN116843013A (en) | Multi-mode data enhancement and coding method for unsupervised contrast learning | |
| CN117154256A (en) | Electrochemical repair method of lithium battery | |
| CN119538977B (en) | Lightweight method, device and equipment for progressive knowledge distillation based on pruning network | |
| CN119360112B (en) | Uniform alignment of pre-trained multimodal model features based on meta-learning | |
| CN105678340B (en) | An Automatic Image Annotation Method Based on Enhanced Stacked Autoencoder | |
| Luu et al. | Clinical name entity recognition based on recurrent neural networks | |
| CN115687989A (en) | Multi-user spectrum sensing method based on Transformer-DNN hybrid network under Laplace noise | |
| Pan et al. | SSER: semi-supervised emotion recognition based on triplet loss and pseudo label | |
| CN115115966A (en) | Video scene segmentation method and device, computer equipment and storage medium |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |