[go: up one dir, main page]

CN114913397A - Modal deficiency-oriented graph representation learning method, system, device and storage medium - Google Patents

Modal deficiency-oriented graph representation learning method, system, device and storage medium Download PDF

Info

Publication number
CN114913397A
CN114913397A CN202210434473.7A CN202210434473A CN114913397A CN 114913397 A CN114913397 A CN 114913397A CN 202210434473 A CN202210434473 A CN 202210434473A CN 114913397 A CN114913397 A CN 114913397A
Authority
CN
China
Prior art keywords
modal
vertex
resource
matrix
super
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210434473.7A
Other languages
Chinese (zh)
Inventor
李明
段嘉峰
梁吉业
王宇光
陈雨婷
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Normal University CJNU
Original Assignee
Zhejiang Normal University CJNU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Normal University CJNU filed Critical Zhejiang Normal University CJNU
Priority to CN202210434473.7A priority Critical patent/CN114913397A/en
Publication of CN114913397A publication Critical patent/CN114913397A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Multimedia (AREA)
  • Image Generation (AREA)

Abstract

The invention discloses a modal deficiency-oriented graph representation learning method, a modal deficiency-oriented graph representation learning system, a modal deficiency-oriented graph representation learning device and a storage medium, and relates to the technical field of computers. According to the method, the characteristic interaction is carried out on the representation matrixes of various modes in the resource hyper-vertex of the modal missing hyper-graph so as to perfect modal data, and then the characteristic fusion matrix of the resource hyper-vertex is obtained, then the characteristic fusion matrix interaction between the resource hyper-vertices is carried out so as to perfect the missing modal data in the resource hyper-vertex, and then the modal aggregation operation is carried out on the resource hyper-vertex after the mode is completed, so that the characteristic extraction vector of the resource hyper-vertex is obtained, the multi-modal hyper-graph is updated, and then the multi-modal hyper-graph is input into the representation model so as to obtain a relatively accurate representation result.

Description

面向模态缺失的图表示学习方法、系统、装置及存储介质Method, system, device and storage medium for graph representation learning oriented to missing modality

技术领域technical field

本发明涉计算机技术领域,尤其涉及一种面向模态缺失的图表示学习方法、系统、装置及存储介质。The present invention relates to the field of computer technology, and in particular, to a method, system, device and storage medium for graph representation learning for modal deletion.

背景技术Background technique

随着数据收集技术的进步,多模态数据急剧增加,多模态数据的整合能够提高机器学习在各种应用场景中的性能,然而如何整合模态缺失的数据仍然是一个具有挑战的问题。目前,大多数的图神经网络都假设顶点的所有特征可以获得,但是实际上顶点的特征可能只有一部分是可以获得的,比如说在社交网络中,部分用户不愿意填写年龄和性别这类信息,造成这类模态数据的缺失。With the advancement of data collection technology, multi-modal data has increased dramatically, and the integration of multi-modal data can improve the performance of machine learning in various application scenarios. However, how to integrate modal-missing data is still a challenging problem. At present, most graph neural networks assume that all the features of the vertices are available, but in fact, only part of the features of the vertices may be available. For example, in social networks, some users are unwilling to fill in information such as age and gender. resulting in the absence of such modal data.

面对模态数据的缺失,目前常采用删除不完整的数据样本或插补缺失的模态等策略来处理这个问题,但是进行数据删除会显著减少训练数据的数量,当具有不同的缺失数据的大规模样本时可能会导致深度学习模型的过度拟合,基于插补的方法会在原始数据中引入额外的噪声,从而对学习模型的性能产生负面影响,并且有时会与复杂的辅助模型相关联,增加了计算的难度。综上,目前对于大规模模态缺失的数据整合方法得到的结果准确度较低。Faced with the lack of modal data, strategies such as deleting incomplete data samples or imputing missing modalities are often used to deal with this problem. However, data deletion will significantly reduce the amount of training data. Large-scale samples can lead to overfitting of deep learning models, imputation-based methods introduce additional noise in the original data, which negatively affects the performance of the learned model, and is sometimes associated with complex auxiliary models , which increases the computational difficulty. To sum up, the accuracy of the results obtained by the current data integration methods for large-scale missing modalities is low.

发明内容SUMMARY OF THE INVENTION

本发明旨在至少解决现有技术中存在的技术问题之一。为此,本发明提出一种面向模态缺失的图表示学习方法、系统、装置及存储介质,能够对模态缺失的图数据进行有效整合,提供较为准确的图表示结果。The present invention aims to solve at least one of the technical problems existing in the prior art. To this end, the present invention proposes a method, system, device and storage medium for modal-missing graph representation learning, which can effectively integrate modal-missing graph data and provide more accurate graph representation results.

一方面,本发明实施例提供了一种面向模态缺失的图表示学习方法,包括以下步骤:On the one hand, an embodiment of the present invention provides a modal deletion-oriented graph representation learning method, including the following steps:

获取模态缺失超图,其中,所述模态缺失超图包括多个资源超顶点,所述资源超顶点包括多种模态的表示矩阵;obtaining a modal missing hypergraph, wherein the modal missing hypergraph includes a plurality of resource hypervertices, and the resource hypervertices include representation matrices of multiple modalities;

对每一个所述资源超顶点内的多种所述模态的表示矩阵进行特征交互,得到每一个所述资源超顶点的特征融合矩阵;performing feature interaction on the representation matrices of a plurality of the modalities in each of the resource super-vertices to obtain a feature fusion matrix of each of the resource super-vertices;

对于每一个资源超顶点,将邻居资源超顶点的特征融合矩阵与当前资源超顶点的特征融合矩阵进行交互融合,并对当前资源超顶点的特征融合矩阵进行模态聚合得到当前资源超顶点的特征提取向量;For each resource super vertex, interactively fuse the feature fusion matrix of the neighbor resource super vertex with the feature fusion matrix of the current resource super vertex, and perform modal aggregation on the feature fusion matrix of the current resource super vertex to obtain the features of the current resource super vertex extract vector;

根据每一个资源超顶点的所述特征提取向量更新所述模态缺失超图,得到多模态超图;Update the missing modal hypergraph according to the feature extraction vector of each resource hypervertex to obtain a multimodal hypergraph;

将所述多模态超图输入图表示模型得到图表示结果。The multimodal hypergraph is input into a graph representation model to obtain a graph representation result.

根据本发明一些实施例,所述图表示模型包括双层自注意力残差连接的图神经网络,所述图神经网络包括基于紧框架构建的滤波器组,所述滤波器组包括低通滤波器、第一高通滤波器和第二高通滤波器。According to some embodiments of the present invention, the graph representation model comprises a two-layer self-attention residual connected graph neural network, the graph neural network comprising a filter bank constructed based on a compact frame, the filter bank comprising low-pass filtering filter, a first high-pass filter, and a second high-pass filter.

根据本发明一些实施例,所述对每一个所述资源超顶点内的多种所述模态的表示矩阵进行特征交互融合,得到每一个所述资源超顶点的特征融合矩阵包括以下步骤:According to some embodiments of the present invention, the feature interaction fusion of the representation matrices of the multiple modalities in each of the resource super-vertices, to obtain the feature fusion matrix of each of the resource super-vertices, includes the following steps:

根据所述资源超顶点内的每一个模态的表示矩阵进行所述模态内的特征融合,得到第一融合向量;Perform feature fusion in the modal according to the representation matrix of each modal in the resource super-vertex to obtain a first fusion vector;

根据所述资源超顶点内的多种所述模态的表示矩阵对所述资源超顶点内的多种所述模态间的特征进行交互,得到第二融合向量;The second fusion vector is obtained by interacting the features between the multiple modes in the resource super vertex according to the representation matrix of the multiple modes in the resource super vertex;

根据多个所述第一融合向量和多个所述第二融合向量得到所述特征融合矩阵,其中,所述特征融合矩阵包括多个模态向量。The feature fusion matrix is obtained according to a plurality of the first fusion vectors and a plurality of the second fusion vectors, wherein the feature fusion matrix includes a plurality of modal vectors.

根据本发明一些实施例,所述根据所述资源超顶点内的每一个模态的表示矩阵进行所述模态内的特征融合,得到第一融合向量包括以下步骤:According to some embodiments of the present invention, performing feature fusion in the modality according to the representation matrix of each modality in the resource super-vertex to obtain a first fusion vector includes the following steps:

根据所述模态的表示矩阵确定所述模态的特征值;determining the eigenvalues of the modality according to the representation matrix of the modality;

对所述模态的表示矩阵中的所有特征向量进行两两内积得到格拉姆矩阵;Perform a pairwise inner product on all eigenvectors in the representation matrix of the modal to obtain a Gram matrix;

根据所述特征值和所述格拉姆矩阵得到所述第一融合向量。The first fusion vector is obtained according to the eigenvalue and the Gram matrix.

根据本发明一些实施例,所述根据多个所述第一融合向量和多个所述第二融合向量得到所述特征融合矩阵包括以下步骤:According to some embodiments of the present invention, obtaining the feature fusion matrix according to a plurality of the first fusion vectors and a plurality of the second fusion vectors includes the following steps:

将所述第一融合向量和所述第二融合向量均输入张量型随机配置神经网络进行向量维度转换和拼接得到所述特征融合矩阵。Inputting the first fusion vector and the second fusion vector into a tensor-type random configuration neural network to perform vector dimension conversion and splicing to obtain the feature fusion matrix.

根据本发明一些实施例,所述将邻居资源超顶点的特征融合矩阵与当前资源超顶点的特征融合矩阵进行交互融合包括以下步骤:According to some embodiments of the present invention, the interactive fusion of the feature fusion matrix of the neighbor resource super-vertex and the feature fusion matrix of the current resource super-vertex includes the following steps:

确定当前资源超顶点的缺失模态;Determine the missing modality of the current resource supervertex;

根据所述邻居资源超顶点对所述缺失模态的第一影响权重和所述邻居资源超顶点的特征融合矩阵中与所述缺失模态对应的模态向量确定所述当前资源超顶点的特征融合矩阵中的缺失模态向量;The feature of the current resource super vertex is determined according to the first influence weight of the neighbor resource super vertex on the missing modality and the modal vector corresponding to the missing modality in the feature fusion matrix of the neighbor resource super vertex the missing modal vectors in the fusion matrix;

根据所述邻居资源超顶点的特征融合矩阵中与所述缺失模态不同的模态向量对所述缺失模态的第二影响权重更新所述当前资源超顶点的特征融合矩阵中的缺失模态向量。Update the missing modality in the feature fusion matrix of the current resource super-vertex according to the second influence weight of the modal vector different from the missing modality in the feature fusion matrix of the neighbor resource super-vertex on the missing modality vector.

根据本发明一些实施例,所述对当前资源超顶点的特征融合矩阵进行模态聚合得到当前资源超顶点的特征提取向量包括以下步骤:According to some embodiments of the present invention, the modal aggregation of the feature fusion matrix of the current resource super vertex to obtain the feature extraction vector of the current resource super vertex includes the following steps:

确定所述特征融合矩阵中多个模态向量的权重系数;determining the weight coefficients of multiple modal vectors in the feature fusion matrix;

将每一个所述模态向量与所述权重系数相乘得到多个加权模态向量;multiplying each of the modal vectors by the weighting coefficient to obtain a plurality of weighted modal vectors;

将多个所述加权模态向量相加得到所述当前资源超顶点的特征提取向量。A feature extraction vector of the current resource super-vertex is obtained by adding up a plurality of the weighted modal vectors.

另一方面,本发明实施例还提供一种面向模态缺失的图表示学习系统,包括:On the other hand, an embodiment of the present invention also provides a modal deletion-oriented graph representation learning system, including:

第一模块,用于获取模态缺失超图,其中,所述模态缺失超图包括多个资源超顶点,所述资源超顶点包括多种模态的表示矩阵;a first module, configured to obtain a modality-missing hypergraph, wherein the modality-missing hypergraph includes multiple resource hypervertices, and the resource hypervertices include representation matrices of multiple modes;

第二模块,用于对每一个所述资源超顶点内的多种所述模态的表示矩阵进行特征交互,得到每一个所述资源超顶点的特征融合矩阵;The second module is used to perform feature interaction on the representation matrices of multiple said modalities in each of the resource super-vertices to obtain a feature fusion matrix of each of the resource super-vertices;

第三模块,用于对于每一个资源超顶点,将邻居资源超顶点的特征融合矩阵与当前资源超顶点的特征融合矩阵进行交互融合,并对当前资源超顶点的特征融合矩阵进行模态聚合得到当前资源超顶点的特征提取向量;The third module is used to interactively fuse the feature fusion matrix of the neighbor resource super vertex with the feature fusion matrix of the current resource super vertex for each resource super vertex, and modally aggregate the feature fusion matrix of the current resource super vertex to obtain The feature extraction vector of the current resource super vertex;

第四模块,用于根据每一个资源超顶点的所述特征提取向量更新所述模态缺失超图,得到多模态超图;The fourth module is used to update the modal missing hypergraph according to the feature extraction vector of each resource hypervertex to obtain a multimodal hypergraph;

第五模块,用于将所述多模态超图输入图表示模型得到图表示结果。The fifth module is used for inputting the multimodal hypergraph into a graph representation model to obtain a graph representation result.

另一方面,本发明实施例还提供一种面向模态缺失的图表示学习装置,包括:On the other hand, an embodiment of the present invention also provides a modal absence-oriented graph representation learning device, including:

至少一个处理器;at least one processor;

至少一个存储器,用于存储至少一个程序;at least one memory for storing at least one program;

当所述至少一个程序被所述至少一个处理器执行,使得至少一个所述处理器实现如前面所述的面向模态缺失的图表示学习方法。When the at least one program is executed by the at least one processor, the at least one processor implements the modal absence-oriented graph representation learning method as described above.

另一方面,本发明实施例还提供一种计算机可读存储介质,所述计算机可读存储介质存储有计算机可执行指令,所述计算机可执行指令用于使计算机执行如前面所述的面向模态缺失的图表示学习方法。On the other hand, an embodiment of the present invention further provides a computer-readable storage medium, where the computer-readable storage medium stores computer-executable instructions, and the computer-executable instructions are used to cause a computer to execute the model-oriented method described above. Graphs with missing states represent learning methods.

本发明上述的技术方案至少具有如下优点或有益效果之一:通过在模态缺失超图的资源超顶点内的多种模态的表示矩阵进行特征交互以完善模态数据,进而得到资源超顶点的特征融合矩阵,然后进行资源超顶点间的特征融合矩阵交互,以完善资源超顶点中的缺失模态的数据,再对模态完善后的资源超顶点的进行模态聚合操作,从而得到资源超顶点的特征提取向量并更新多模态超图,然后将多模态超图输入图表示模型得到较为准确的图表示结果。The above technical solution of the present invention has at least one of the following advantages or beneficial effects: by performing feature interaction between the representation matrices of multiple modalities in the resource hypervertex of the modal-missing hypergraph to improve the modal data, and then obtain the resource hypervertex feature fusion matrix, and then perform the feature fusion matrix interaction between resource super-vertices to improve the missing modal data in the resource super-vertices, and then perform modal aggregation operations on the resource super-vertices after modal improvement, so as to obtain resources The feature extraction vector of the super vertices and update the multi-modal hypergraph, and then input the multi-modal hypergraph into the graph representation model to obtain a more accurate graph representation result.

附图说明Description of drawings

图1是本发明实施例提供的面向模态缺失的图表示学习方法流程图;FIG. 1 is a flowchart of a modal deletion-oriented graph representation learning method provided by an embodiment of the present invention;

图2是本发明实施例提供的面向模态缺失的图表示学习系统示意图;2 is a schematic diagram of a modal deletion-oriented graph representation learning system provided by an embodiment of the present invention;

图3是本发明实施例提供的面向模态缺失的图表示学习装置示意图;3 is a schematic diagram of a modal deletion-oriented graph representation learning device provided by an embodiment of the present invention;

图4是本发明实施例提供的模态缺失超图构建过程示意图;4 is a schematic diagram of a process of constructing a modal deletion hypergraph provided by an embodiment of the present invention;

图5是本发明实施例提供的资源超顶点特征表示与内部特征交互过程示意图;5 is a schematic diagram of an interaction process between resource super-vertex feature representation and internal features provided by an embodiment of the present invention;

图6是本发明实施例提供的模态缺失超图在三层图注意力网络机制的模型处理过程示意图;6 is a schematic diagram of a model processing process of a modal-missing hypergraph in a three-layer graph attention network mechanism provided by an embodiment of the present invention;

图7是本发明实施例提供的基于紧框架的图神经网络构建过程示意图;7 is a schematic diagram of the construction process of a graph neural network based on a tight frame provided by an embodiment of the present invention;

图8是本发明实施例提供的多模态超图在基于双层自注意力残差连接的图神经网络的图表示模型的处理过程示意图。FIG. 8 is a schematic diagram of a processing process of a graph representation model of a multi-modal hypergraph in a graph neural network based on two-layer self-attention residual connections provided by an embodiment of the present invention.

具体实施方式Detailed ways

下面详细描述本发明的实施例,所述实施例的示例在附图中示出,其中自始至终相同或者类似的标号表示相同或者类似的原件或具有相同或类似功能的元件。下面通过参考附图描述的实施例是示例性的,仅用于解释本发明,而不能理解为对本发明的限制。The following describes in detail the embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein the same or similar reference numerals refer to the same or similar elements or elements having the same or similar functions throughout. The embodiments described below with reference to the accompanying drawings are exemplary, only used to explain the present invention, and should not be construed as a limitation of the present invention.

在本发明的描述中,需要理解的是,涉及到方位描述,例如上、下、左、右等指示的方位或者位置关系为基于附图所示的方位或者位置关系,仅是为了便于描述本发明和简化描述,而不是指示或者暗示所指的装置或元件必须具有特定的方位、以特定的方位构造和操作,因此不能理解为对本发明的限制。In the description of the present invention, it should be understood that the azimuth description, for example, the azimuth or position relationship indicated by up, down, left, right, etc., is based on the azimuth or position relationship shown in the accompanying drawings, which is only for the convenience of describing the present invention. The invention and simplified description do not indicate or imply that the device or element referred to must have a particular orientation, be constructed and operate in a particular orientation, and therefore should not be construed as limiting the invention.

本发明的描述中,如果有描述到第一、第二等只是用于区分技术特征为目的,而不能理解为指示或者暗示相对重要性或者隐含指明所指示的技术特征的数量或者隐含指明所指示的技术特征的先后关系。In the description of the present invention, if the first, second, etc. are described only for the purpose of distinguishing technical features, it should not be understood as indicating or implying relative importance or implicitly indicating the number of indicated technical features or implicitly indicating The order of the indicated technical features.

本发明实施例提供了一种面向模态缺失的图表示学习方法,参照图1,本发明实施例的面向模态缺失的图表示学习方法包括但不限于步骤S110、步骤S120、步骤S130、步骤S140和步骤S150。An embodiment of the present invention provides a modal absence-oriented graph representation learning method. Referring to FIG. 1 , the modal absence oriented graph representation learning method according to an embodiment of the present invention includes but is not limited to step S110 , step S120 , step S130 , and step S130 . S140 and step S150.

步骤S110,获取模态缺失超图,其中,模态缺失超图包括多个资源超顶点,资源超顶点包括多种模态的表示矩阵;Step S110, acquiring a modality-missing hypergraph, wherein the modality-missing hypergraph includes multiple resource hypervertices, and the resource hypervertex includes representation matrices of multiple modes;

在一些实施例中,在现实的社交网络中,用户和资源之间互动产生数据,但是并不是对于所有的用户都会产生完整模态的资源,可能存在用户模态缺失的情况,基于此,构建的图结构为模态缺失超图。示例性地,参照图4,定义三种模态数据分别为i1、i2、i3,用户u1产生模态i1的资源,用户u2产生模态i1、i2的资源,用户u3产生模态i1、i2、i3的资源,用户u4产生模态i2、i3的资源,因此,构建对应于用户的资源超顶点,并根据用户间关系得到模态缺失图,资源超顶点

Figure BDA0003612436020000041
包括用户u1包括模态i1的数据,资源超顶点
Figure BDA0003612436020000042
包括用户u2和模态i1、i2的数据,资源超顶点
Figure BDA0003612436020000058
包括用户u3和模态i1、i2、i3的数据,资源超顶点
Figure BDA0003612436020000059
用户u4和模态i2、i3的数据。为了有效地表示超顶点之间的高阶连接,使用超边E共享资源超顶点的相似信息并显示模态缺失超图中的资源超顶点之间的关系,根据资源超顶点和超边构建包含多个资源超顶点的模态缺失超图G。In some embodiments, in a real social network, the interaction between users and resources generates data, but not all users will generate resources with complete modalities, and there may be cases where user modalities are missing. Based on this, construct The graph structure is a modal-missing hypergraph. Exemplarily, referring to FIG. 4 , three modal data are defined as i 1 , i 2 , and i 3 respectively. User u 1 generates resources of modal i 1 , and user u 2 generates resources of modal i 1 , i 2 , User u 3 generates resources of modalities i 1 , i 2 , i 3 , and user u 4 generates resources of modal i 2 , i 3 . Therefore, the resource super-vertex corresponding to the user is constructed, and the modalities are obtained according to the relationship between users Missing graph, resource hypervertex
Figure BDA0003612436020000041
include data for user u 1 including modality i 1 , resource supervertex
Figure BDA0003612436020000042
Contains data for user u 2 and modalities i 1 , i 2 , resource hypervertex
Figure BDA0003612436020000058
Contains data for user u 3 and modalities i 1 , i 2 , i 3 , resource hypervertex
Figure BDA0003612436020000059
Data for user u 4 and modalities i 2 , i 3 . In order to efficiently represent higher-order connections between hypervertices, hyperedges E are used to share similar information of resource hypervertices and show the relationship between resource hypervertices in the modal-missing hypergraph. Modal missing hypergraph G for multiple resource hypervertices.

进一步地,参照图5,对模态缺失超图中的单一模态数据进行表示,利用简单的MLP(多层感知机)来构建编码器,其中各模态对应的编码器参数不共享,从而各模态的表示矩阵。定义fm(·;θm)为模态m的MLP的嵌入网络,θm为可训练学习的参数,对于每个资源超顶点vi的模态m的表示矩阵如公式(1)所示:Further, referring to FIG. 5 , the single modal data in the modal absence hypermap is represented, and a simple MLP (Multilayer Perceptron) is used to construct an encoder, wherein the encoder parameters corresponding to each modal are not shared, so that Representation matrix for each mode. Define f m (·; θ m ) as the MLP embedding network of modality m, θ m is a parameter that can be trained and learned, and the representation matrix of modality m for each resource super-vertex v i is shown in formula (1) :

Figure BDA0003612436020000051
Figure BDA0003612436020000051

其中,

Figure BDA0003612436020000052
Fm为模态m的嵌入维度。in,
Figure BDA0003612436020000052
F m is the embedding dimension of modality m.

步骤S120,对每一个资源超顶点内的多种模态的表示矩阵进行特征交互,得到每一个资源超顶点的特征融合矩阵;Step S120, performing feature interaction on the representation matrices of multiple modalities in each resource super-vertex to obtain a feature fusion matrix of each resource super-vertex;

在一些实施例中,在得到模态的表示矩阵之后,为了缩小资源超顶点内部各种模态抽象表示之间的“语义鸿沟”,需在资源超顶点内部实现单一模态特征交互编码及各模态间特征新编码的深度交互。In some embodiments, after obtaining the representation matrix of the modalities, in order to narrow the "semantic gap" between the abstract representations of various modalities within the resource super-vertex, it is necessary to implement a single-modal feature interactive encoding and various modalities within the resource super-vertex. Deep interactions for new encodings of features between modalities.

具体地,参照图5,首先,对于单一模态的表示矩阵,使用Gram矩阵计算特征之间存在的关系,并分别得到单一模态对应的第一融合向量

Figure BDA0003612436020000053
Gram矩阵即格拉姆矩阵,n维欧式空间中任意k个向量y1,y2,...yk之间两两的内积所组成的矩阵,即为这k个向量的Gram矩阵,Gram矩阵是一个对称矩阵,Gram矩阵用于度量矩阵中各个维度本身的特性以及各个维度之间的关系,矩阵中的特征向量内积之后得到的多尺度矩阵中,对角线元素提供了不同特征向量本身的信息,其余元素提供了不同特征向量之间的相关信息。采用Gram矩阵变换的第一融合向量既能体现出有特征的个数,又能体现出不同特征间的紧密程度,其主要的表示形式如公式(2)所示:Specifically, referring to FIG. 5, first, for the representation matrix of a single modality, use the Gram matrix to calculate the relationship between the features, and obtain the first fusion vector corresponding to the single modality respectively
Figure BDA0003612436020000053
Gram matrix is Gram matrix, the matrix formed by the inner product of any k vectors y 1 , y 2 , ... y k in n-dimensional Euclidean space, is the Gram matrix of these k vectors, Gram The matrix is a symmetric matrix. The Gram matrix is used to measure the characteristics of each dimension in the matrix and the relationship between each dimension. In the multi-scale matrix obtained after the inner product of the eigenvectors in the matrix, the diagonal elements provide different eigenvectors. information of itself, and the remaining elements provide relevant information between different feature vectors. The first fusion vector transformed by Gram matrix can not only reflect the number of features, but also reflect the degree of closeness between different features. Its main representation is shown in formula (2):

Figure BDA0003612436020000054
Figure BDA0003612436020000054

其次,对于资源超顶点中每对模态或者两个以上的模态之间可能会出现高阶模态的相互作用,对于不同模态的表示矩阵,通过模态间表示矩阵两两外积的方式可以得到表示两者之间信息交互的二阶张量

Figure BDA0003612436020000055
再通过计算三个不同模态的表示矩阵之间的外积得到表示三者信息交互的三阶张量
Figure BDA0003612436020000056
以此类推即得到有限个不同阶次的张量,通过各个模态的表示矩阵之间的外积进行深度交互,得到多个第二融合向量
Figure BDA0003612436020000057
Secondly, for each pair of modalities or between more than two modalities in the resource super-vertex, there may be higher-order modal interactions. For the representation matrices of different modalities, the outer product of the representation matrices between the modalities can be used. Get a second-order tensor representing the information interaction between the two
Figure BDA0003612436020000055
Then, by calculating the outer product between the representation matrices of the three different modes, a third-order tensor representing the information interaction of the three is obtained.
Figure BDA0003612436020000056
By analogy, a limited number of tensors of different orders are obtained, and through the deep interaction between the representation matrices of each mode, a plurality of second fusion vectors are obtained.
Figure BDA0003612436020000057

进一步地,假设H(·)表示幂集运算并且对于每个子集

Figure BDA0003612436020000061
可以从每个因子S上学到一种多模态互动及与其相关的一系列的信息。资源超顶点内的特征互补信息由许多因子组成,每个因子用一个F’维向量表示,在进行资源超顶点内的各模态特征交互的计算过程如下:Further, assume that H( ) represents a power set operation and for each subset
Figure BDA0003612436020000061
A multimodal interaction and a series of information related to it can be learned from each factor S. The feature complementary information in the resource super-vertex is composed of many factors, and each factor is represented by an F'-dimensional vector. The calculation process of each modal feature interaction in the resource super-vertex is as follows:

如果S中只有一个元素,表示正在计算某一种模态m∈S的第一融合向量,第一融合向量

Figure BDA0003612436020000062
的计算过程如公式组(3)所示:If there is only one element in S, it means that the first fusion vector of a certain mode m∈S is being calculated, the first fusion vector
Figure BDA0003612436020000062
The calculation process is shown in formula group (3):

Figure BDA0003612436020000063
Figure BDA0003612436020000063

Figure BDA0003612436020000064
Figure BDA0003612436020000064

Figure BDA0003612436020000065
Figure BDA0003612436020000065

其中,

Figure BDA0003612436020000066
bm∈RF′,
Figure BDA0003612436020000067
bm,m∈RF′,均为神经网络gm(·)和gm,m(·)的可训练参数,
Figure BDA0003612436020000068
为表示矩阵
Figure BDA0003612436020000069
对应的Gram矩阵,
Figure BDA00036124360200000610
表示模态m的特征值,该特征值可以为模态m的特征均值。in,
Figure BDA0003612436020000066
b m ∈ R F ′,
Figure BDA0003612436020000067
b m, m ∈ R F ′, both are trainable parameters of neural networks g m ( ) and g m, m ( ),
Figure BDA0003612436020000068
to represent the matrix
Figure BDA0003612436020000069
The corresponding Gram matrix,
Figure BDA00036124360200000610
Represents the eigenvalue of mode m, which can be the eigenvalue of mode m.

如果S中存在超过一个的元素,表示正在计算不同模态

Figure BDA00036124360200000611
之间的跨模态交互信息的第二融合向量。第二融合向量
Figure BDA00036124360200000612
的计算过程如公式组(4)所示:If there is more than one element in S, it means that different modes are being calculated
Figure BDA00036124360200000611
A second fusion vector of cross-modal interaction information between. second fusion vector
Figure BDA00036124360200000612
The calculation process is shown in formula group (4):

Figure BDA00036124360200000613
Figure BDA00036124360200000613

Figure BDA00036124360200000614
Figure BDA00036124360200000614

其中,首先由

Figure BDA00036124360200000615
求得S中的单一模态m的表示矩阵
Figure BDA00036124360200000616
的交叉验证,进而得到S中相关模态表示矩阵的|S|层交叉验证
Figure BDA00036124360200000617
最后再将
Figure BDA00036124360200000618
输入图神经网络gS(·)中进行训练,bS和Us均为图神经网络gS(·)的可训练参数。Among them, the first
Figure BDA00036124360200000615
Find the representation matrix for a single mode m in S
Figure BDA00036124360200000616
, and then obtain the |S| layer cross-validation of the relevant modal representation matrix in S
Figure BDA00036124360200000617
Finally, the
Figure BDA00036124360200000618
The input graph neural network g S ( ) is used for training, and b S and U s are both trainable parameters of the graph neural network g S ( ).

在获得第一融合向量和第二融合向量之后,利用支持张量输入的随机配置神经网络实现不同阶次张量到相同维度向量表示的转换。通过上述的过程,得到维度不同、阶次不同的多个第一融合向量和多个第二融合向量,因此,将多个第一融合向量和多个第二融合向量均输入支持张量输入的随机配置神经网络得到统一维度表示的融合向量

Figure BDA00036124360200000619
在本发明实施例中的图5,可以得到7种融合向量,分别为
Figure BDA00036124360200000620
根据多种融合向量得到资源超顶点的进行模态特征交互后的特征融合矩阵,其中,特征融合矩阵包括多个模态向量。After obtaining the first fusion vector and the second fusion vector, a random configuration neural network supporting tensor input is used to realize the conversion of tensors of different orders to vector representations of the same dimension. Through the above process, a plurality of first fusion vectors and a plurality of second fusion vectors with different dimensions and different orders are obtained. Therefore, the plurality of first fusion vectors and the plurality of second fusion vectors are input into the support tensor input Randomly configure the neural network to obtain a fusion vector represented by a uniform dimension
Figure BDA00036124360200000619
In FIG. 5 in the embodiment of the present invention, 7 kinds of fusion vectors can be obtained, which are
Figure BDA00036124360200000620
A feature fusion matrix of the resource super-vertex after modal feature interaction is obtained according to various fusion vectors, wherein the feature fusion matrix includes a plurality of modal vectors.

步骤S130,对于每一个资源超顶点,将邻居资源超顶点的特征融合矩阵与当前资源超顶点的特征融合矩阵进行交互融合,并对当前资源超顶点的特征融合矩阵进行模态聚合得到当前资源超顶点的特征提取向量;Step S130, for each resource super vertex, interactively fuse the feature fusion matrix of the neighbor resource super vertex and the feature fusion matrix of the current resource super vertex, and perform modal aggregation on the feature fusion matrix of the current resource super vertex to obtain the current resource super vertex. Feature extraction vector of vertices;

在一些实施例中,由于部分资源超顶点内模态缺失会引起资源超顶点之间的模态信息不均衡,因此可以设计基于三层图注意力网络机制的模型补全和完整资源超顶点中缺少的模态数据。对于基于三层图注意力网络机制的模型的构建,在第一层图注意力网络中训练具有与缺失模态同一模态的邻居超顶点对当前资源超顶点的缺失模态的第一影响权重,在第二层图注意力网络中训练与缺失模态不同模态的邻居资源超顶点对对当前资源超顶点的缺失模态的,第二影响权重,在第三层图注意力网络中训练资源超顶点中不同模态的权重系数。通过上述的基于三层图注意力网络机制的模型可以基于与缺失模态相同的邻居资源超顶点的模态对当前资源超顶点的缺失模态进行补全,再基于与缺失模态相同的邻居资源超顶点的模态对当前资源超顶点的缺失模态进行完善,然后对资源超顶点中的多种模态进行聚合以得到资源超顶点的特征提取向量。In some embodiments, since the lack of modalities in some resource supervertices will cause the modal information imbalance between resource supervertices, model completion and complete resource supervertices based on the three-layer graph attention network mechanism can be designed. Missing modal data. For the construction of the model based on the three-layer graph attention network mechanism, the first influence weight of the missing modality of the current resource supervertex on the neighbor supervertices with the same modality as the missing modality is trained in the first layer graph attention network , trained in the second-layer graph attention network for neighbor resource supervertices of different modalities from the missing modality on the missing modality of the current resource supervertex, the second influence weights, trained in the third-layer graph attention network Weight coefficients for different modalities in the resource supervertex. Through the above-mentioned model based on the three-layer graph attention network mechanism, the missing modalities of the current resource super-vertex can be completed based on the modalities of the same neighbor resource super-vertices as the missing modalities, and then based on the same neighbors as the missing modalities. The modality of the resource supervertex completes the missing modality of the current resource supervertex, and then aggregates multiple modes in the resource supervertex to obtain the feature extraction vector of the resource supervertex.

示例性地,以资源超顶点之间为例,说明注意力网络的求解过程。Exemplarily, the solution process of the attention network is described by taking the resource super-vertices as an example.

模态缺失超图经过张量型随机配置网络得到的数据为h={h1,h2,h3,...hN},hi∈RF,N代表资源超顶点的数量,F表示资源超顶点中的特征数量,矩阵h的大小是N×F,矩阵h表示所有资源超顶点的特征,R表示某一个资源超顶点的特征,R的大小为F×1。经过注意力网络的输出数据可以表示为{h1′,h2′,h3′,...hN′},hi′∈RF,F′表示资源超顶点的输出特征向量维度,设置权值矩阵W表示注意力网络输入与输出特征之间的关系。The data obtained by the modal-missing hypergraph through a tensor-type random configuration network is h={h 1 , h 2 , h 3 ,...h N } , hi ∈ R F , N represents the number of resource super vertices, F Represents the number of features in the resource super-vertex, the size of the matrix h is N×F, the matrix h represents the features of all resource super-vertices, R represents the feature of a resource super-vertex, and the size of R is F×1. The output data through the attention network can be expressed as {h 1 ', h 2 ', h 3 ', ... h N ' } , hi '∈RF , F ' represents the output feature vector dimension of the resource super-vertex, The weight matrix W is set to represent the relationship between the input and output features of the attention network.

针对每个资源超顶点实行自注意的机制,则资源超顶点vj对于资源超顶点vi的重要性如公式(5)所示,其中关系函数定义为a:The self-attention mechanism is implemented for each resource super vertex, then the importance of the resource super vertex v j to the resource super vertex v i is shown in formula (5), where the relation function is defined as a:

eij=a(Whi,Whj); (5)e ij =a(Wh i , Wh j ); (5)

为了使注意力系数更容易计算和便于比较,引入了softmax函数对所有与资源超顶点vi相邻的资源超顶点vj进行正则化得到注意力系数,注意力系数αij计算如公式(6)所示:In order to make the attention coefficient easier to calculate and compare, a softmax function is introduced to regularize all resource super-vertices v j adjacent to resource super-vertices v i to obtain the attention coefficients. The attention coefficient α ij is calculated as formula (6 ) as shown:

Figure BDA0003612436020000071
Figure BDA0003612436020000071

上述注意力系数经过注意力网络的输出层的LeakyReLU函数后,得到的完整注意力系数如下公式(7)所示,其中,a是注意力网络中连接层与层之间的权重矩阵。After the above attention coefficient passes through the LeakyReLU function of the output layer of the attention network, the complete attention coefficient obtained is shown in the following formula (7), where a is the weight matrix between the connection layers in the attention network.

Figure BDA0003612436020000072
Figure BDA0003612436020000072

进一步地,根据上述资源超顶点间的注意力系数可以预测每个资源超顶点的输出特征如公式(8)所示:Further, according to the above-mentioned attention coefficient between resource super-vertices, the output feature of each resource super-vertex can be predicted as shown in formula (8):

Figure BDA0003612436020000073
Figure BDA0003612436020000073

在本实施例中,将上述注意力网络求解的方法应用到本实施例的三层图注意力网络机制的模型中,从而实现资源超顶点的缺失模态的补全完善和资源超顶点内的模态聚合,得到资源超顶点的特征提取向量。In this embodiment, the above-mentioned method for solving the attention network is applied to the model of the three-layer graph attention network mechanism in this embodiment, so as to realize the completion and improvement of the missing modalities of the resource super-vertex and the Modal aggregation to obtain feature extraction vectors of resource supervertices.

参照图6,首先,将模态缺失超图输入第一层自注意网络中,得到资源超顶点vi的所有邻居顶点对于资源超顶点vi的缺失模态m的表达式如公式(9)所示:Referring to Fig. 6, first, the modal-missing hypergraph is input into the first-layer self-attention network, and the expression of the missing modality m of the resource super-vertex v i for all the neighbor vertices of the resource super-vertex v i is obtained as formula (9) shown:

Figure BDA0003612436020000081
Figure BDA0003612436020000081

其中,σ为激活函数,

Figure BDA0003612436020000082
为资源超顶点vi的所有邻居资源超顶点vj对于资源超顶点vi的缺失模态m的注意力系数,
Figure BDA0003612436020000083
表示邻居资源超顶点vj的模态向量m,Wmn表示模态向量m到映射低维的模态向量n的可训练权重,
Figure BDA0003612436020000084
表示资源超顶点vi的邻居资源超顶点,
Figure BDA0003612436020000085
的计算如公式(10)所示:where σ is the activation function,
Figure BDA0003612436020000082
is the attention coefficient of resource super-vertex v j for all neighbors of resource super-vertex v i for the missing modality m of resource super-vertex v i ,
Figure BDA0003612436020000083
represents the modal vector m of the neighbor resource super-vertex v j , W mn represents the trainable weight of the modal vector m to the modal vector n of the mapping low-dimensional,
Figure BDA0003612436020000084
represents the neighbor resource supervertex of resource supervertex v i ,
Figure BDA0003612436020000085
The calculation of is shown in formula (10):

Figure BDA0003612436020000086
Figure BDA0003612436020000086

其中,αm表示资源超顶点vj对于资源超顶点vi的第一影响权重。Among them, α m represents the first influence weight of the resource super-vertex v j to the resource super-vertex v i .

通过上述过程,可以得到资源超顶点vi增加有信息

Figure BDA0003612436020000087
k是资源超顶点vi所具有的邻居的数量。Through the above process, it can be obtained that the resource super vertex v i has added information
Figure BDA0003612436020000087
k is the number of neighbors that the resource supervertex vi has.

其次,将模态缺失超图输入第二层自注意网络中,得到资源超顶点vi的所有邻居顶点完善资源超顶点vi的缺失模态m,得到资源超顶点vi的模态m的表达式如公式(11)所示:Second, input the modal-missing hypergraph into the second-layer self-attention network, get all the neighbor vertices of the resource super-vertex v i to complete the missing modal m of the resource-super-vertex v i , and get the modal m of the resource super-vertex v i . The expression is shown in formula (11):

Figure BDA0003612436020000088
Figure BDA0003612436020000088

其中,σ为激活函数,βm,n为模态m与模态n之间的注意力系数,

Figure BDA0003612436020000089
表示资源超顶点vi的低维度模态向量n,
Figure BDA00036124360200000810
表示模态向量n到模态向量m空间的转换,βm,n的计算如公式(12)所示:where σ is the activation function, β m, n is the attention coefficient between modality m and modality n,
Figure BDA0003612436020000089
represents the low-dimensional modal vector n of the resource supervertex v i ,
Figure BDA00036124360200000810
Represents the transformation of the modal vector n to the modal vector m space, and the calculation of β m, n is shown in formula (12):

Figure BDA00036124360200000811
Figure BDA00036124360200000811

其中,βm,n表示邻居超顶点的其他模态对缺失模态的第二影响权重。where βm ,n represents the second influence weight of other modalities of neighboring supervertices on the missing modality.

通过上述过程,将资源超顶点vi的所有模态信息补充完整。Through the above process, all the modal information of the resource super vertex v i is completed.

然后,将模态缺失超图输入第三层自注意网络中对每个资源超顶点进行模态融合,得到资源超顶点vi的特征提取矩阵hi′,hi′计算如公式(13)所示:Then, the modal-missing hypergraph is input into the third-layer self-attention network to perform modal fusion on each resource super-vertex to obtain the feature extraction matrix h i ′ of the resource super-vertex v i , and h i ′ is calculated as formula (13) shown:

Figure BDA00036124360200000812
Figure BDA00036124360200000812

其中,wj表示资源超顶点vi内模态向量j和其他模态向量的权重系数,

Figure BDA00036124360200000813
表示资源超顶点vi内模态向量j,N表示资源超顶点vi内的模态向量数量。Among them, w j represents the weight coefficient of the modal vector j and other modal vectors in the resource super-vertex v i ,
Figure BDA00036124360200000813
represents the modal vector j in the resource super-vertex v i , and N represents the number of modal vectors in the resource super-vertex v i .

步骤S140,根据每一个资源超顶点的特征提取向量更新模态缺失超图,得到多模态超图;Step S140, updating the modal missing hypergraph according to the feature extraction vector of each resource hypervertex to obtain a multimodal hypergraph;

步骤S150,将多模态超图输入图表示模型得到图表示结果。Step S150, inputting the multimodal hypergraph into the graph representation model to obtain a graph representation result.

在一些实施例中,图表示模型包括双层自注意力残差连接的图神经网络,图神经网络包括基于紧框架构建的滤波器组,滤波器组包括低通滤波器、第一高通滤波器和第二高通滤波器。In some embodiments, the graph representation model includes a two-layer self-attention residual connected graph neural network, the graph neural network includes a filter bank constructed based on a compact frame, the filter bank includes a low-pass filter, a first high-pass filter and a second high pass filter.

其中,参照图7,图神经网络通过以下方式构建:Among them, referring to Figure 7, the graph neural network is constructed in the following manner:

针对得到的大规模的模态缺失超图Γ,通过多层聚类对图的特征进行提取。将大规模的模态缺失超图细化为不同层次的子图

Figure BDA00036124360200000814
以此得到树形结构。For the obtained large-scale modal-missing hypergraph Γ, the features of the graph are extracted through multi-layer clustering. Refinement of large-scale modal-missing hypergraphs into subgraphs at different levels
Figure BDA00036124360200000814
This results in a tree structure.

根据树形结构在树的每一层构造对应该层的正交基。正交基的好处在于可以很方便地表示空间中的点,用数学语言表示为两个向量u和v的乘积,当u·v=0,这两个向量相互垂直或相互正交,u和v也即该层的正交基,同时,层与层之间可以快速实现正交基共享与扩充,以此得到图上的多尺度紧框架变换,该变换具有等价于快速傅里叶变换(FFT)的算法复杂度。According to the tree structure, an orthogonal basis corresponding to the layer is constructed at each layer of the tree. The advantage of the orthonormal basis is that it can easily represent a point in the space, which is expressed as the product of two vectors u and v in mathematical language. When u·v=0, these two vectors are perpendicular or orthogonal to each other, and u and v is also the orthonormal basis of the layer. At the same time, the sharing and expansion of the orthonormal basis can be quickly realized between the layers, so as to obtain the multi-scale compact frame transform on the graph, which is equivalent to the fast Fourier transform. (FFT) algorithmic complexity.

基于卷积定理设计基于多尺度紧框架的图卷积网络,该网络运用基于张量的小波变换,直接把傅里叶变换F(ω)的基替换,即将无限长的三角函数基换成了有限长的会衰减的小波基,不仅能够获取信号的频率,还可以定位到时间,解决了傅里叶变换不能刻画时间域上信号的局部特性以及没有时频分析的缺点,小波变换函数WT(α,τ)的表示如公式(14)所示,其中,小波变换有尺度α和平移量τ两个变量。尺度α控制小波函数的伸缩,平移量τ控制小波函数的平移,f(t)表示输入的图信号,

Figure BDA0003612436020000091
表示小波的基。Based on the convolution theorem, a graph convolution network based on a multi-scale compact frame is designed. The network uses a tensor-based wavelet transform to directly replace the basis of the Fourier transform F(ω), that is, the infinitely long trigonometric function basis is replaced by The finite wavelet base with attenuation can not only obtain the frequency of the signal, but also locate the time, which solves the shortcomings that the Fourier transform cannot describe the local characteristics of the signal in the time domain and has no time-frequency analysis. The wavelet transform function WT ( The expression of α, τ) is shown in formula (14), wherein, the wavelet transform has two variables: scale α and translation τ. The scale α controls the expansion and contraction of the wavelet function, the translation amount τ controls the translation of the wavelet function, f(t) represents the input graph signal,
Figure BDA0003612436020000091
represents the basis of the wavelet.

Figure BDA0003612436020000092
Figure BDA0003612436020000092

设计图上的紧框架变换,具体的框架操作流程为:通过对给定一个具有结构(邻接矩阵)和特征信息图的图信号f,构造一个低通和两个高通小波变换矩阵Wr,j|r=0,1,...,n;j=1,...,j},同时,定义

Figure BDA0003612436020000093
是r次切比雪夫多项式,
Figure BDA0003612436020000094
是图拉普拉斯矩阵,将其与输入特征矩阵连续相乘以产生小波系数,可表示为公式(15):To design the compact frame transformation on the graph, the specific frame operation flow is: by giving a graph signal f with a structure (adjacency matrix) and a feature information graph, construct a low-pass and two high-pass wavelet transform matrices W r, j |r=0,1,...,n; j=1,...,j}, meanwhile, define
Figure BDA0003612436020000093
is a Chebyshev polynomial of degree r,
Figure BDA0003612436020000094
is the graph Laplacian matrix, which is continuously multiplied by the input feature matrix to generate wavelet coefficients, which can be expressed as Equation (15):

Figure BDA0003612436020000095
Figure BDA0003612436020000095

图信号可以根据建立的框架进行分解,然后用对偶框架重构,同时调整正交基集里一些新的向量以及空间位置,并且设置上下界相等,成为紧框架。把输入的图信号与基或框架作内积以进行函数空间到系数空间的变换,变换后的能量(内积的平方和度量)仍然有一个大于0的上下界,由于框架的冗余性,框架内低通和两个高通小波变换函数系数的表达不具有唯一性。此外,系数通过可训练的网络滤波器进行计算,并通过压缩率进行图像的压缩,得到图卷积正确的嵌入图,收缩激活的帧卷积表示如公式(16)所示:The graph signal can be decomposed according to the established frame, and then reconstructed with the dual frame, while adjusting some new vectors and spatial positions in the orthogonal basis set, and setting the upper and lower bounds equal to become a compact frame. The input graph signal is inner product with the basis or frame to transform the function space to the coefficient space, and the transformed energy (the measure of the square sum of the inner product) still has an upper and lower bound greater than 0. Due to the redundancy of the frame, The expression of the low-pass and two high-pass wavelet transform function coefficients within the framework is not unique. In addition, the coefficients are calculated by the trainable network filter, and the image is compressed by the compression rate, and the correct embedded graph of the graph convolution is obtained. The frame convolution representation of the contraction activation is shown in formula (16):

λShrinkage(diag(θ)(WX′))),X′=XW; (16)λShrinkage(diag(θ)(WX'))), X'=XW; (16)

其中,θ为网络滤波器,X是图

Figure BDA0003612436020000096
的特征矩阵,重构算子λ是分解算子W的帧变换矩阵的重新排列。通过再次使用转置对齐的变换矩阵,重构激活系数并将其发送回空间域作为卷积输出。where θ is the network filter and X is the graph
Figure BDA0003612436020000096
The feature matrix of , the reconstruction operator λ is the rearrangement of the frame transformation matrix of the decomposition operator W. By using the transposed aligned transformation matrix again, the activation coefficients are reconstructed and sent back to the spatial domain as the convolution output.

在本实施例中,参照图8,多模态超图在基于双层自注意力残差连接的图神经网络的图表示模型中的处理过程为:In this embodiment, referring to FIG. 8 , the processing procedure of the multimodal hypergraph in the graph representation model of the graph neural network based on the two-layer self-attention residual connection is as follows:

对于多模态超图的特征提取向量输入上述多尺度框架的图神经网络进行正常的滤波操作,得到的输出向量为x(1),将原来输入的特征提取向量不经过加工直接到输出层得到向量x(2),将经过分解矩阵后得到的向量记为x(3),将向量x(1)、x(2)和x(3)经过加权平均后得到输出向量x′,输出向量x′通过公式(17)计算:For the feature extraction vector of the multi-modal hypergraph, input it into the graph neural network of the above multi-scale framework to perform normal filtering operations, and the obtained output vector is x (1) . The vector x (2) , the vector obtained after decomposing the matrix is denoted as x (3) , and the vector x (1) , x (2) and x (3) are weighted and averaged to obtain the output vector x', the output vector x ' is calculated by formula (17):

Figure BDA0003612436020000101
Figure BDA0003612436020000101

其中,i取值为1,2,3,wi表示x(i)的权重。Among them, i takes the value of 1, 2, 3, and w i represents the weight of x (i) .

经过多次重复上述步骤,然后经过重构因子λ得到最后的图表示结果x,从而实现对大规模的模态缺失超图的图表示学习。After repeating the above steps many times, the final graph representation result x is obtained through the reconstruction factor λ, so as to realize the graph representation learning of large-scale modal-missing hypergraphs.

在本实施例中,拟融合小波分析的方法来设计图上的紧框架变换,以此定义新的谱图卷积,并设计快速卷积计算算法进行特征提取向量的卷积运算。通过构建基于残差连接的深度递归图神经网络实现高效图表示学习。基于上述紧框架的图卷积神经网络,支持高频、低频图信息的协同处理,即分别通过高通滤波(high-pass)和低通滤波(low-pass)双通道分层进行图上的资源超顶点信息的聚合,相对于现有的图神经网络只支持低通滤波,本实施例的图神经网络能够避免低通滤波造成的信息损失。In this embodiment, the method of wavelet analysis is proposed to be combined to design a tight frame transformation on the graph, thereby defining a new spectral graph convolution, and a fast convolution calculation algorithm is designed to perform the convolution operation of the feature extraction vector. Efficient graph representation learning is achieved by building a deep recurrent graph neural network based on residual connections. The graph convolutional neural network based on the above-mentioned compact framework supports the collaborative processing of high-frequency and low-frequency graph information, that is, the resources on the graph are layered through high-pass and low-pass filters respectively. The aggregation of super-vertex information, compared with the existing graph neural network that only supports low-pass filtering, the graph neural network of this embodiment can avoid information loss caused by low-pass filtering.

根据本发明一些具体实施例,步骤S120包括但不限于以下步骤:According to some specific embodiments of the present invention, step S120 includes but is not limited to the following steps:

步骤S210,根据资源超顶点内的每一个模态的表示矩阵进行模态内的特征融合,得到第一融合向量;Step S210, perform feature fusion in the modality according to the representation matrix of each modality in the resource super-vertex to obtain a first fusion vector;

步骤S220,根据资源超顶点内的多种模态的表示矩阵对资源超顶点内的多种模态间的特征进行交互,得到第二融合向量;Step S220, according to the representation matrix of the various modalities in the resource super-vertex, interact with the features between the various modalities in the resource super-vertex to obtain a second fusion vector;

步骤S230,根据多个第一融合向量和多个第二融合向量得到特征融合矩阵,其中,特征融合矩阵包括多个模态向量。Step S230, obtaining a feature fusion matrix according to the plurality of first fusion vectors and the plurality of second fusion vectors, wherein the feature fusion matrix includes a plurality of modal vectors.

根据本发明一些具体实施例,步骤S210包括但不限于以下步骤:According to some specific embodiments of the present invention, step S210 includes but is not limited to the following steps:

步骤S310,根据模态的表示矩阵确定模态的特征值;Step S310, determining the eigenvalues of the modality according to the representation matrix of the modality;

步骤S320,对模态的表示矩阵中的所有特征向量进行两两内积得到格拉姆矩阵;Step S320, performing a pairwise inner product on all eigenvectors in the modal representation matrix to obtain a Gram matrix;

步骤S330,根据特征值和所述格拉姆矩阵得到第一融合向量。Step S330, obtaining a first fusion vector according to the eigenvalue and the Gram matrix.

根据本发明一些具体实施例,步骤S230包括但不限于以下步骤:According to some specific embodiments of the present invention, step S230 includes but is not limited to the following steps:

步骤S410,将第一融合向量和第二融合向量均输入张量型随机配置神经网络进行向量维度转换和拼接得到特征融合矩阵。Step S410, inputting the first fusion vector and the second fusion vector into a tensor-type random configuration neural network to perform vector dimension transformation and splicing to obtain a feature fusion matrix.

根据本发明一些具体实施例,步骤S130中的将邻居资源超顶点的特征融合矩阵与当前资源超顶点的特征融合矩阵进行交互融合包括但不限于以下步骤:According to some specific embodiments of the present invention, the interactive fusion of the feature fusion matrix of the neighbor resource super-vertex and the feature fusion matrix of the current resource super-vertex in step S130 includes but is not limited to the following steps:

步骤S510,确定当前资源超顶点的缺失模态;Step S510, determining the missing mode of the current resource super vertex;

步骤S520,根据邻居资源超顶点对缺失模态的第一影响权重和邻居资源超顶点的特征融合矩阵中与缺失模态对应的模态向量确定当前资源超顶点的特征融合矩阵中的缺失模态向量;Step S520, determining the missing mode in the feature fusion matrix of the current resource super-vertex according to the first influence weight of the neighbor resource super-vertex on the missing mode and the modal vector corresponding to the missing mode in the feature fusion matrix of the neighbor resource super-vertex vector;

步骤S530,根据邻居资源超顶点的特征融合矩阵中与缺失模态不同的模态向量对缺失模态的第二影响权重更新当前资源超顶点的特征融合矩阵中的缺失模态向量。Step S530: Update the missing modal vector in the feature fusion matrix of the current resource super-vertex according to the second influence weight of the modal vector different from the missing modal in the feature fusion matrix of the neighboring resource super-vertex on the missing modal.

根据本发明一些具体实施例,步骤S130中的对当前资源超顶点的特征融合矩阵进行模态聚合得到当前资源超顶点的特征提取向量包括但不限于以下步骤:According to some specific embodiments of the present invention, performing modal aggregation on the feature fusion matrix of the current resource super-vertex to obtain the feature extraction vector of the current resource super-vertex in step S130 includes but is not limited to the following steps:

步骤S610,确定特征融合矩阵中多个模态向量的权重系数;Step S610, determining the weight coefficients of multiple modal vectors in the feature fusion matrix;

步骤S620,将每一个模态向量与权重系数相乘得到多个加权模态向量;Step S620, multiply each modal vector by the weight coefficient to obtain a plurality of weighted modal vectors;

步骤S630,将多个加权模态向量相加得到当前资源超顶点的特征提取向量。Step S630, adding a plurality of weighted modal vectors to obtain a feature extraction vector of the current resource super vertex.

另一方面,本发明实施例还提供一种面向模态缺失的图表示学习系统,参照图2,面向模态缺失的图表示学习系统包括:On the other hand, an embodiment of the present invention also provides a modal absence-oriented graph representation learning system. Referring to FIG. 2 , the modal absence oriented graph representation learning system includes:

第一模块,用于获取模态缺失超图,其中,所述模态缺失超图包括多个资源超顶点,所述资源超顶点包括多种模态的表示矩阵;a first module, configured to obtain a modality-missing hypergraph, wherein the modality-missing hypergraph includes multiple resource hypervertices, and the resource hypervertices include representation matrices of multiple modes;

第二模块,用于对每一个所述资源超顶点内的多种所述模态的表示矩阵进行特征交互,得到每一个所述资源超顶点的特征融合矩阵;The second module is used to perform feature interaction on the representation matrices of multiple said modalities in each of the resource super-vertices to obtain a feature fusion matrix of each of the resource super-vertices;

第三模块,用于对于每一个资源超顶点,将邻居资源超顶点的特征融合矩阵与当前资源超顶点的特征融合矩阵进行交互融合,并对当前资源超顶点的特征融合矩阵进行模态聚合得到当前资源超顶点的特征提取向量;The third module is used to interactively fuse the feature fusion matrix of the neighbor resource super vertex with the feature fusion matrix of the current resource super vertex for each resource super vertex, and modally aggregate the feature fusion matrix of the current resource super vertex to obtain The feature extraction vector of the current resource super vertex;

第四模块,用于根据每一个资源超顶点的所述特征提取向量更新所述模态缺失超图,得到多模态超图;The fourth module is used to update the modal missing hypergraph according to the feature extraction vector of each resource hypervertex to obtain a multimodal hypergraph;

第五模块,用于将所述多模态超图输入图表示模型得到图表示结果。The fifth module is used for inputting the multimodal hypergraph into a graph representation model to obtain a graph representation result.

可以理解的是,上述面向模态缺失的图表示学习方法实施例中的内容均适用于本系统实施例中,本系统实施例所具体实现的功能与上述面向模态缺失的图表示学习方法实施例相同,并且达到的有益效果与上述面向模态缺失的图表示学习方法实施例所达到的有益效果也相同。It can be understood that, the contents in the above-mentioned embodiments of the learning method for modal absence-oriented graph representation are all applicable to the embodiments of this system, and the functions specifically implemented by the embodiments of this system are the same as the implementation of the above-mentioned modal absence-oriented graph representation learning method. Examples are the same, and the beneficial effects achieved are also the same as the beneficial effects achieved by the above-mentioned embodiment of the modal deletion-oriented graph representation learning method.

参照图3,图3是本发明一个实施例提供的面向模态缺失的图表示学习装置的示意图。本发明实施例的面向模态缺失的图表示学习装置包括一个或多个控制处理器和存储器,图3中以一个控制处理器及一个存储器为例。Referring to FIG. 3 , FIG. 3 is a schematic diagram of a modal deletion-oriented graph representation learning apparatus provided by an embodiment of the present invention. The modal deletion-oriented graph representation learning device according to the embodiment of the present invention includes one or more control processors and a memory. In FIG. 3 , one control processor and one memory are used as an example.

控制处理器和存储器可以通过总线或者其他方式连接,图3中以通过总线连接为例。The control processor and the memory may be connected by a bus or in other ways, and the connection by a bus is taken as an example in FIG. 3 .

存储器作为一种非暂态计算机可读存储介质,可用于存储非暂态软件程序以及非暂态性计算机可执行程序。此外,存储器可以包括高速随机存取存储器,还可以包括非暂态存储器,例如至少一个磁盘存储器件、闪存器件、或其他非暂态固态存储器件。在一些实施方式中,存储器可选包括相对于控制处理器远程设置的存储器,这些远程存储器可以通过网络连接至该面向模态缺失的图表示学习装置。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。As a non-transitory computer-readable storage medium, the memory can be used to store non-transitory software programs and non-transitory computer-executable programs. Additionally, the memory may include high-speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid state storage device. In some embodiments, the memory may optionally include memory located remotely from the control processor, and these remote memories may be connected to the modality-absence-oriented graphical representation learning device via a network. Examples of such networks include, but are not limited to, the Internet, an intranet, a local area network, a mobile communication network, and combinations thereof.

本领域技术人员可以理解,图3中示出的装置结构并不构成对面向模态缺失的图表示学习装置的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置。Those skilled in the art can understand that the device structure shown in FIG. 3 does not constitute a limitation on the modal absence-oriented graphical representation learning device, and may include more or less components than those shown in the figure, or combine some components, Or a different component arrangement.

实现上述实施例中应用于面向模态缺失的图表示学习装置的面向模态缺失的图表示学习方法所需的非暂态软件程序以及指令存储在存储器中,当被控制处理器执行时,执行上述实施例中应用于面向模态缺失的图表示学习装置的面向模态缺失的图表示学习方法。The non-transitory software programs and instructions required to realize the modal absence oriented graph representation learning method applied to the modal absence oriented graph representation learning device in the above embodiment are stored in the memory, and when executed by the control processor, execute In the above-mentioned embodiments, the modal absence-oriented graph representation learning method applied to the modal absence-oriented graph representation learning apparatus.

此外,本发明的一个实施例还提供了一种计算机可读存储介质,该计算机可读存储介质存储有计算机可执行指令,该计算机可执行指令被一个或多个控制处理器执行,可使得上述一个或多个控制处理器执行上述方法实施例中的面向模态缺失的图表示学习方法。In addition, an embodiment of the present invention also provides a computer-readable storage medium, where the computer-readable storage medium stores computer-executable instructions, and the computer-executable instructions are executed by one or more control processors, so that the above-mentioned One or more control processors execute the modality-absent-oriented graph representation learning method in the above method embodiments.

本领域普通技术人员可以理解,上文中所公开方法中的全部或某些步骤、系统可以被实施为软件、固件、硬件及其适当的组合。某些物理组件或所有物理组件可以被实施为由处理器,如中央处理器、数字信号处理器或微处理器执行的软件,或者被实施为硬件,或者被实施为集成电路,如专用集成电路。这样的软件可以分布在计算机可读介质上,计算机可读介质可以包括计算机存储介质(或非暂时性介质)和通信介质(或暂时性介质)。如本领域普通技术人员公知的,术语计算机存储介质包括在用于存储信息(诸如计算机可读指令、数据结构、程序模块或其他数据)的任何方法或技术中实施的易失性和非易失性、可移除和不可移除介质。计算机存储介质包括但不限于RAM、ROM、EEPROM、闪存或其他存储器技术、CD-ROM、数字多功能盘(DVD)或其他光盘存储、磁盒、磁带、磁盘存储或其他磁存储装置、或者可以用于存储期望的信息并且可以被计算机访问的任何其他的介质。此外,本领域普通技术人员公知的是,通信介质通常包含计算机可读指令、数据结构、程序模块或者诸如载波或其他传输机制之类的调制数据信号中的其他数据,并且可包括任何信息递送介质。Those of ordinary skill in the art can understand that all or some of the steps and systems in the methods disclosed above can be implemented as software, firmware, hardware, and appropriate combinations thereof. Some or all physical components may be implemented as software executed by a processor, such as a central processing unit, digital signal processor or microprocessor, or as hardware, or as an integrated circuit, such as an application specific integrated circuit . Such software may be distributed on computer-readable media, which may include computer storage media (or non-transitory media) and communication media (or transitory media). As known to those of ordinary skill in the art, the term computer storage media includes both volatile and nonvolatile implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules or other data flexible, removable and non-removable media. Computer storage media include, but are not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disk (DVD) or other optical disk storage, magnetic cartridges, magnetic tape, magnetic disk storage or other magnetic storage devices, or may Any other medium used to store desired information and which can be accessed by a computer. In addition, communication media typically embodies computer readable instructions, data structures, program modules, or other data in a modulated data signal such as a carrier wave or other transport mechanism, and can include any information delivery media, as is well known to those of ordinary skill in the art .

上面结合附图对本发明实施例作了详细说明,但是本发明不限于上述实施例,在所属技术领域普通技术人员所具备的知识范围内,还可以在不脱离本发明宗旨的前提下作出各种变化。The embodiments of the present invention have been described in detail above in conjunction with the accompanying drawings, but the present invention is not limited to the above-mentioned embodiments, and within the scope of knowledge possessed by those of ordinary skill in the art, various Variety.

Claims (10)

1.一种面向模态缺失的图表示学习方法,其特征在于,包括以下步骤:1. A method for learning representations of modal absences, comprising the following steps: 获取模态缺失超图,其中,所述模态缺失超图包括多个资源超顶点,所述资源超顶点包括多种模态的表示矩阵;obtaining a modal missing hypergraph, wherein the modal missing hypergraph includes a plurality of resource hypervertices, and the resource hypervertices include representation matrices of multiple modalities; 对每一个所述资源超顶点内的多种所述模态的表示矩阵进行特征交互,得到每一个所述资源超顶点的特征融合矩阵;performing feature interaction on the representation matrices of a plurality of the modalities in each of the resource super-vertices to obtain a feature fusion matrix of each of the resource super-vertices; 对于每一个资源超顶点,将邻居资源超顶点的特征融合矩阵与当前资源超顶点的特征融合矩阵进行交互融合,并对当前资源超顶点的特征融合矩阵进行模态聚合得到当前资源超顶点的特征提取向量;For each resource super vertex, interactively fuse the feature fusion matrix of the neighbor resource super vertex with the feature fusion matrix of the current resource super vertex, and perform modal aggregation on the feature fusion matrix of the current resource super vertex to obtain the features of the current resource super vertex extract vector; 根据每一个资源超顶点的所述特征提取向量更新所述模态缺失超图,得到多模态超图;Update the missing modal hypergraph according to the feature extraction vector of each resource hypervertex to obtain a multimodal hypergraph; 将所述多模态超图输入图表示模型得到图表示结果。The multimodal hypergraph is input into a graph representation model to obtain a graph representation result. 2.根据权利要求1所述的面向模态缺失的图表示学习方法,其特征在于,所述图表示模型包括双层自注意力残差连接的图神经网络,所述图神经网络包括基于紧框架构建的滤波器组,所述滤波器组包括低通滤波器、第一高通滤波器和第二高通滤波器。2 . The modal deletion-oriented graph representation learning method according to claim 1 , wherein the graph representation model comprises a graph neural network with two-layer self-attention residual connections, and the graph neural network comprises a graph neural network based on tight A filter bank constructed by the framework, the filter bank includes a low-pass filter, a first high-pass filter, and a second high-pass filter. 3.根据权利要求1所述的面向模态缺失的图表示学习方法,其特征在于,所述对每一个所述资源超顶点内的多种所述模态的表示矩阵进行特征交互融合,得到每一个所述资源超顶点的特征融合矩阵包括以下步骤:3. The modal deletion-oriented graph representation learning method according to claim 1, wherein the feature interaction fusion is performed on the representation matrices of a plurality of the modalities in each of the resource super-vertices to obtain The feature fusion matrix of each resource super-vertex includes the following steps: 根据所述资源超顶点内的每一个模态的表示矩阵进行所述模态内的特征融合,得到第一融合向量;Perform feature fusion in the modal according to the representation matrix of each modal in the resource super-vertex to obtain a first fusion vector; 根据所述资源超顶点内的多种所述模态的表示矩阵对所述资源超顶点内的多种所述模态间的特征进行交互,得到第二融合向量;The second fusion vector is obtained by interacting the features between the multiple modes in the resource super vertex according to the representation matrix of the multiple modes in the resource super vertex; 根据多个所述第一融合向量和多个所述第二融合向量得到所述特征融合矩阵,其中,所述特征融合矩阵包括多个模态向量。The feature fusion matrix is obtained according to a plurality of the first fusion vectors and a plurality of the second fusion vectors, wherein the feature fusion matrix includes a plurality of modal vectors. 4.根据权利要求3所述的面向模态缺失的图表示学习方法,其特征在于,所述根据所述资源超顶点内的每一个模态的表示矩阵进行所述模态内的特征融合,得到第一融合向量包括以下步骤:4. The modal absence-oriented graph representation learning method according to claim 3, wherein the feature fusion in the modality is performed according to the representation matrix of each modality in the resource super-vertex, Obtaining the first fusion vector includes the following steps: 根据所述模态的表示矩阵确定所述模态的特征值;determining the eigenvalues of the modality according to the representation matrix of the modality; 对所述模态的表示矩阵中的所有特征向量进行两两内积得到格拉姆矩阵;Perform a pairwise inner product on all eigenvectors in the representation matrix of the modal to obtain a Gram matrix; 根据所述特征值和所述格拉姆矩阵得到所述第一融合向量。The first fusion vector is obtained according to the eigenvalue and the Gram matrix. 5.根据权利要求3所述的面向模态缺失的图表示学习方法,其特征在于,所述根据多个所述第一融合向量和多个所述第二融合向量得到所述特征融合矩阵包括以下步骤:5 . The modal deletion-oriented graph representation learning method according to claim 3 , wherein the obtaining the feature fusion matrix according to a plurality of the first fusion vectors and a plurality of the second fusion vectors comprises the following steps: 6 . The following steps: 将所述第一融合向量和所述第二融合向量均输入张量型随机配置神经网络进行向量维度转换和拼接得到所述特征融合矩阵。Inputting the first fusion vector and the second fusion vector into a tensor-type random configuration neural network to perform vector dimension conversion and splicing to obtain the feature fusion matrix. 6.根据权利要求5所述的面向模态缺失的图表示学习方法,其特征在于,所述将邻居资源超顶点的特征融合矩阵与当前资源超顶点的特征融合矩阵进行交互融合包括以下步骤:6. The modal deletion-oriented graph representation learning method according to claim 5, characterized in that, the described feature fusion matrix of neighbor resource super-vertex and the feature fusion matrix of current resource super-vertex are interactively fused and comprise the following steps: 确定当前资源超顶点的缺失模态;Determine the missing modality of the current resource supervertex; 根据所述邻居资源超顶点对所述缺失模态的第一影响权重和所述邻居资源超顶点的特征融合矩阵中与所述缺失模态对应的模态向量确定所述当前资源超顶点的特征融合矩阵中的缺失模态向量;The feature of the current resource super vertex is determined according to the first influence weight of the neighbor resource super vertex on the missing modality and the modal vector corresponding to the missing modality in the feature fusion matrix of the neighbor resource super vertex the missing modal vectors in the fusion matrix; 根据所述邻居资源超顶点的特征融合矩阵中与所述缺失模态不同的模态向量对所述缺失模态的第二影响权重更新所述当前资源超顶点的特征融合矩阵中的缺失模态向量。Update the missing modality in the feature fusion matrix of the current resource super-vertex according to the second influence weight of the modal vector different from the missing modality in the feature fusion matrix of the neighbor resource super-vertex on the missing modality vector. 7.根据权利要求6所述的面向模态缺失的图表示学习方法,其特征在于,所述对当前资源超顶点的特征融合矩阵进行模态聚合得到当前资源超顶点的特征提取向量包括以下步骤:7. The modal absence-oriented graph representation learning method according to claim 6, wherein the feature extraction vector that the feature fusion matrix of the current resource super-vertex is obtained by modal aggregation to obtain the current resource super-vertex comprises the following steps : 确定所述特征融合矩阵中多个模态向量的权重系数;determining the weight coefficients of multiple modal vectors in the feature fusion matrix; 将每一个所述模态向量与所述权重系数相乘得到多个加权模态向量;multiplying each of the modal vectors by the weighting coefficient to obtain a plurality of weighted modal vectors; 将多个所述加权模态向量相加得到所述当前资源超顶点的特征提取向量。A feature extraction vector of the current resource super-vertex is obtained by adding up a plurality of the weighted modal vectors. 8.一种面向模态缺失的图表示学习系统,其特征在于,包括:8. A modal deletion-oriented graph representation learning system, characterized in that it comprises: 第一模块,用于获取模态缺失超图,其中,所述模态缺失超图包括多个资源超顶点,所述资源超顶点包括多种模态的表示矩阵;a first module, configured to obtain a modality-missing hypergraph, wherein the modality-missing hypergraph includes multiple resource hypervertices, and the resource hypervertices include representation matrices of multiple modes; 第二模块,用于对每一个所述资源超顶点内的多种所述模态的表示矩阵进行特征交互,得到每一个所述资源超顶点的特征融合矩阵;The second module is used to perform feature interaction on the representation matrices of multiple said modalities in each of the resource super-vertices to obtain a feature fusion matrix of each of the resource super-vertices; 第三模块,用于对于每一个资源超顶点,将邻居资源超顶点的特征融合矩阵与当前资源超顶点的特征融合矩阵进行交互融合,并对当前资源超顶点的特征融合矩阵进行模态聚合得到当前资源超顶点的特征提取向量;The third module is used to interactively fuse the feature fusion matrix of the neighbor resource super vertex with the feature fusion matrix of the current resource super vertex for each resource super vertex, and modally aggregate the feature fusion matrix of the current resource super vertex to obtain The feature extraction vector of the current resource super vertex; 第四模块,用于根据每一个资源超顶点的所述特征提取向量更新所述模态缺失超图,得到多模态超图;The fourth module is used to update the modal missing hypergraph according to the feature extraction vector of each resource hypervertex to obtain a multimodal hypergraph; 第五模块,用于将所述多模态超图输入图表示模型得到图表示结果。The fifth module is used for inputting the multimodal hypergraph into a graph representation model to obtain a graph representation result. 9.一种面向模态缺失的图表示学习装置,其特征在于,包括:9. A modal absence-oriented graph representation learning device, characterized in that it comprises: 至少一个处理器;at least one processor; 至少一个存储器,用于存储至少一个程序;at least one memory for storing at least one program; 当所述至少一个程序被所述至少一个处理器执行,使得至少一个所述处理器实现如权利要求1至7任一项所述的面向模态缺失的图表示学习方法。When the at least one program is executed by the at least one processor, the at least one processor implements the modal absence-oriented graph representation learning method according to any one of claims 1 to 7. 10.一种计算机可读存储介质,其中存储有处理器可执行的程序,其特征在于,所述处理器可执行的程序被由所述处理器执行时用于实现如权利要求1至7任一项所述的面向模态缺失的图表示学习方法。10. A computer-readable storage medium in which a processor-executable program is stored, wherein the processor-executable program is used to implement any of claims 1 to 7 when executed by the processor. A described modality-missing-oriented graph representation learning method.
CN202210434473.7A 2022-04-24 2022-04-24 Modal deficiency-oriented graph representation learning method, system, device and storage medium Pending CN114913397A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210434473.7A CN114913397A (en) 2022-04-24 2022-04-24 Modal deficiency-oriented graph representation learning method, system, device and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210434473.7A CN114913397A (en) 2022-04-24 2022-04-24 Modal deficiency-oriented graph representation learning method, system, device and storage medium

Publications (1)

Publication Number Publication Date
CN114913397A true CN114913397A (en) 2022-08-16

Family

ID=82764701

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210434473.7A Pending CN114913397A (en) 2022-04-24 2022-04-24 Modal deficiency-oriented graph representation learning method, system, device and storage medium

Country Status (1)

Country Link
CN (1) CN114913397A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115761436A (en) * 2022-11-08 2023-03-07 齐鲁工业大学 Multimodal resource balancing method and system based on associative memory neural network

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106095951A (en) * 2016-06-13 2016-11-09 哈尔滨工程大学 Data space multi-dimensional indexing method based on load balancing and inquiry log
US20180197050A1 (en) * 2016-12-30 2018-07-12 Hrl Laboratories, Llc Zero-shot learning using multi-scale manifold alignment
CN112151071A (en) * 2020-09-23 2020-12-29 哈尔滨工程大学 A speech emotion recognition method based on hybrid wavelet packet feature deep learning
CN114357124A (en) * 2022-03-18 2022-04-15 成都考拉悠然科技有限公司 Video paragraph positioning method based on language reconstruction and graph mechanism
CN114372573A (en) * 2022-01-07 2022-04-19 中国人民解放军国防科技大学 User portrait information recognition method and device, computer equipment and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106095951A (en) * 2016-06-13 2016-11-09 哈尔滨工程大学 Data space multi-dimensional indexing method based on load balancing and inquiry log
US20180197050A1 (en) * 2016-12-30 2018-07-12 Hrl Laboratories, Llc Zero-shot learning using multi-scale manifold alignment
CN112151071A (en) * 2020-09-23 2020-12-29 哈尔滨工程大学 A speech emotion recognition method based on hybrid wavelet packet feature deep learning
CN114372573A (en) * 2022-01-07 2022-04-19 中国人民解放军国防科技大学 User portrait information recognition method and device, computer equipment and storage medium
CN114357124A (en) * 2022-03-18 2022-04-15 成都考拉悠然科技有限公司 Video paragraph positioning method based on language reconstruction and graph mechanism

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
FENG JI: "GFCN: A New Graph Convolutional Network Based on Parallel Flows", ARXIV, 25 February 2019 (2019-02-25), pages 1 - 25 *
JIAYI CHEN: "HGMF: Heterogeneous Graph-based Fusion for Multimodal Data with Incompleteness", KDD \'20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 20 August 2020 (2020-08-20), pages 1295, XP058987192, DOI: 10.1145/3394486.3403182 *
XUEBIN ZHENG: "How Framelets Enhance Graph Neural Networks", ARXIV, 13 February 2021 (2021-02-13), pages 1 - 24 *
李钝;李伦;张行进;郑志蕴;曹付元;: "一种基于结构和属性的图聚类算法研究", 小型微型计算机系统, no. 07, 15 July 2016 (2016-07-15), pages 1469 - 1473 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115761436A (en) * 2022-11-08 2023-03-07 齐鲁工业大学 Multimodal resource balancing method and system based on associative memory neural network

Similar Documents

Publication Publication Date Title
CN109033095B (en) Target transformation method based on attention mechanism
US11836572B2 (en) Quantum inspired convolutional kernels for convolutional neural networks
EP4481587A1 (en) Structure data generation method and apparatus, device, medium, and program product
CN111091010B (en) Similarity determination, network training, search method and device and storage medium
CN111797589A (en) A text processing network, a method for training a neural network, and related equipment
WO2021253671A1 (en) Magnetic resonance cine imaging method and apparatus, and imaging device and storage medium
CN117788949A (en) Image classification method based on personalized federated learning and conditional generative adversarial network
CN112464057A (en) Network data classification method, device, equipment and readable storage medium
CN111316290A (en) System and method for interactive representation learning transfer via deep learning of feature ontology
CN115115723A (en) Image reconstruction model generation method, image reconstruction device, image reconstruction equipment and medium
CN115081615A (en) Neural network training method, data processing method and equipment
CN113065997B (en) Image processing method, neural network training method and related equipment
CN119537705B (en) A multimodal recommendation method and system based on spectral polynomial filter
Ostonov et al. Rlss: A deep reinforcement learning algorithm for sequential scene generation
CN114913397A (en) Modal deficiency-oriented graph representation learning method, system, device and storage medium
CN114331827B (en) Style migration method, device, equipment and storage medium
CN117689966A (en) Quantum Bayesian neural network-based magnetic resonance image classification method
CN117235533B (en) Object variable analysis method, device, computer equipment and storage medium
CN117271808A (en) A multimedia recommendation method based on information bottleneck denoising
CN113901247B (en) Method for detecting optical image target and computing equipment
EP4302236B1 (en) Image processing
CN116503320A (en) Hyperspectral image anomaly detection method, device, equipment and readable storage medium
CN119206245A (en) Method executed by electronic device, storage medium, and program product
CN114419249A (en) Object three-dimensional shape reconstruction method, device, equipment and storage medium
CN114626284A (en) Model processing method and related device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination