CN116740300B - Multi-mode-based prime body and texture fusion furniture model reconstruction method - Google Patents
Multi-mode-based prime body and texture fusion furniture model reconstruction method Download PDFInfo
- Publication number
- CN116740300B CN116740300B CN202310714863.4A CN202310714863A CN116740300B CN 116740300 B CN116740300 B CN 116740300B CN 202310714863 A CN202310714863 A CN 202310714863A CN 116740300 B CN116740300 B CN 116740300B
- Authority
- CN
- China
- Prior art keywords
- texture
- vector
- furniture
- point
- feature vector
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/10—Constructive solid geometry [CSG] using solid primitives, e.g. cylinders, cubes
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
- G06N3/0455—Auto-encoder networks; Encoder-decoder networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/04—Texture mapping
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/08—Volume rendering
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/761—Proximity, similarity or dissimilarity measures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/806—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02P—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
- Y02P90/00—Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
- Y02P90/30—Computing systems specially adapted for manufacturing
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computational Linguistics (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Computer Graphics (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Geometry (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Multimedia (AREA)
- Image Generation (AREA)
Abstract
本发明公开了一种基于多模态的素体与纹理融合家具模型重建方法。该方法包括:获取家具稀疏点云和纹理嵌入向量;将所述家具稀疏点云和所述纹理嵌入向量输入素体与纹理融合模型,得到新纹理家具模型。本发明能够使用户在不需要专业家具设计人员的情况下,完成家具模型的纹理更新,大大节省了时间,并且原始数据只用提供二维图片数据,用户可以使用普通相机获得。
The present invention discloses a furniture model reconstruction method based on multi-modal body and texture fusion. The method comprises: obtaining furniture sparse point cloud and texture embedding vector; inputting the furniture sparse point cloud and the texture embedding vector into a body and texture fusion model to obtain a new texture furniture model. The present invention enables users to complete the texture update of furniture models without the need for professional furniture designers, greatly saving time, and the original data only needs to provide two-dimensional picture data, which users can obtain using ordinary cameras.
Description
技术领域Technical Field
本发明属于基于多模态的三维重建技术领域,尤其涉及到一种基于多模态的素体与纹理融合家具模型重建方法。The present invention belongs to the technical field of three-dimensional reconstruction based on multi-modality, and in particular relates to a furniture model reconstruction method based on multi-modality body and texture fusion.
背景技术Background technique
随着互联网技术的发展及家具行业智能化需求的日渐增加,家具商家可以通过网络媒介展示各种家具效果图,而不需要客户亲自前往现场查看家具实物,这大大节省了时间,方便了商家与客户,在一定程度上带动了家具行业的效率。但是,一个家具模型,需要专业的家具设计人员使用如blender、3ds max和C4D等家具建模软件花费大量时间来完成,大大限制了家具模型推出的效率,因此商家和用户需要一种快速且精准的方法来实现家具模型重建。With the development of Internet technology and the increasing demand for intelligent furniture industry, furniture merchants can display various furniture renderings through online media without the need for customers to go to the site to view the real furniture in person, which greatly saves time, facilitates merchants and customers, and to a certain extent drives the efficiency of the furniture industry. However, a furniture model requires professional furniture designers to spend a lot of time using furniture modeling software such as blender, 3ds max and C4D to complete, which greatly limits the efficiency of furniture model launch. Therefore, merchants and users need a fast and accurate method to achieve furniture model reconstruction.
传统计算机视觉的方法中,通常使用多个2D图像或视频来构建3D模型。例如,可以使用多视角几何的方法,将多个图像或视频帧中的像素坐标转换为3D坐标点云,然后使用这些点云数据生成家具模型。使用深度学习的计算机视觉方法中,通常用神经网络模型来从2D图像或点云数据中预测家具的三维模型。例如,NeRF神经网络体渲染技术,能达到不错的效果,但该模型训练时间太长,且缺乏泛化能力。Instant-NGP神经网络创造性地引入hash编码到神经网络中,提升了模型推理的速度,但其导出的三维模型表面纹理效果不佳且模型消耗显存过大,使得NeRF神经网络在家具领域仍旧没有很好的应用。In traditional computer vision methods, multiple 2D images or videos are usually used to build a 3D model. For example, the pixel coordinates in multiple images or video frames can be converted into 3D coordinate point clouds using multi-view geometry methods, and then these point cloud data can be used to generate furniture models. In computer vision methods using deep learning, neural network models are usually used to predict the three-dimensional model of furniture from 2D images or point cloud data. For example, NeRF neural network volume rendering technology can achieve good results, but the model training time is too long and lacks generalization ability. The Instant-NGP neural network creatively introduces hash coding into the neural network, which improves the speed of model reasoning, but the surface texture of the exported three-dimensional model is not good and the model consumes too much video memory, so the NeRF neural network is still not well applied in the furniture field.
发明内容Summary of the invention
为了解决上述提出的至少一个技术问题,本发明提供一种基于多模态的素体与纹理融合家具模型重建方法,以解决商家和用户需要的一种快速且精准的方法来实现家具模型重建这一技术问题。In order to solve at least one of the technical problems mentioned above, the present invention provides a furniture model reconstruction method based on multimodal body and texture fusion to solve the technical problem of merchants and users needing a fast and accurate method to achieve furniture model reconstruction.
一种基于多模态的素体与纹理融合家具模型重建方法,包括:A furniture model reconstruction method based on multi-modal body and texture fusion, comprising:
获取家具稀疏点云和纹理嵌入向量;Get furniture sparse point cloud and texture embedding vector;
将所述家具稀疏点云和所述纹理嵌入向量输入素体与纹理融合模型,得到新纹理家具模型。The furniture sparse point cloud and the texture embedding vector are input into the body and texture fusion model to obtain a new texture furniture model.
优选地,所述素体与纹理融合模型包括:Preferably, the body and texture fusion model includes:
素体编码器,用于根据家具稀疏点云生成空间坐标点潜在向量;The body encoder is used to generate the spatial coordinate point potential vector based on the furniture sparse point cloud;
素体解码器,用于根据空间坐标点潜在向量生成素体点特征向量和UV坐标特征向量;A voxel decoder is used to generate a voxel point feature vector and a UV coordinate feature vector according to a spatial coordinate point potential vector;
纹理编码器,用于根据纹理嵌入向量生成纹理特征向量;A texture encoder, for generating a texture feature vector according to the texture embedding vector;
特征融合模块,用于根据纹理特征向量和UV坐标特征向量生成UV贴图特征向量,并根据UV贴图特征向量和素体点特征向量生成新纹理家具模型。The feature fusion module is used to generate a UV mapping feature vector according to the texture feature vector and the UV coordinate feature vector, and to generate a new texture furniture model according to the UV mapping feature vector and the body point feature vector.
优选地,将所述家具稀疏点云和纹理嵌入向量输入素体与纹理融合模型,得到新纹理家具模型包括:Preferably, the furniture sparse point cloud and texture embedding vector are input into the body and texture fusion model to obtain a new texture furniture model, which includes:
将所述家具稀疏点云输入素体编码器,得到空间坐标点潜在向量;Input the furniture sparse point cloud into a body encoder to obtain a potential vector of a spatial coordinate point;
将所述空间坐标点潜在向量输入素体解码器,得到素体点特征向量和UV坐标特征向量;Input the spatial coordinate point potential vector into a body decoder to obtain a body point feature vector and a UV coordinate feature vector;
将所述纹理嵌入向量输入纹理编码器,得到纹理特征向量;Inputting the texture embedding vector into a texture encoder to obtain a texture feature vector;
将所述纹理特征向量和所述UV坐标特征向量输入特征融合模块,得到UV贴图特征向量;Input the texture feature vector and the UV coordinate feature vector into a feature fusion module to obtain a UV mapping feature vector;
将所述UV贴图特征向量与所述素体点特征向量相加,经过卷积网络,使用open3D拓展库,得到新纹理家具模型。The UV map feature vector is added to the body point feature vector, and a new texture furniture model is obtained by passing through a convolutional network and using the open3D extension library.
优选地,将所述纹理嵌入向量输入纹理编码器,输出纹理特征向量,包括:Preferably, the texture embedding vector is input into a texture encoder, and a texture feature vector is output, comprising:
将有位置信息的纹理嵌入向量,经过层归一化后将层归一化结果和多头注意力计算结果残差连接,得到纹理潜在向量:The texture with position information is embedded in the vector. After layer normalization, the layer normalization result and the multi-head attention calculation result are residually connected to obtain the texture potential vector:
E‘patch=MSA(LN(Epatch)+Epatch E'patch =MSA(LN( Epatch )+ Epatch
其中E‘patch是纹理潜在向量,Epatch是有位置信息的纹理嵌入向量;Where E'patch is the texture potential vector, and Epatch is the texture embedding vector with position information;
所述纹理潜在向量经过层归一化,全连接层,后将层归一化和全连接层计算结果做残差计算,输出纹理特征向量:The texture potential vector is normalized and fully connected layer, and then the residual calculation is performed on the layer normalization and fully connected layer calculation results to output the texture feature vector:
T=MLP(LN(E‘patch))+E‘patch T=MLP(LN( E'patch ))+ E'patch
其中T是纹理特征向量,E‘patch是纹理潜在向量。Where T is the texture feature vector and E'patch is the texture latent vector.
优选地,将所述空间坐标点潜在向量输入素体解码器,得到素体点特征向量和UV坐标特征向量,包括:Preferably, the spatial coordinate point potential vector is input into a body decoder to obtain a body point feature vector and a UV coordinate feature vector, including:
将空间坐标点潜在向量输入素体解码器,找到空间中任意一点的邻居节点潜在向量集合,其中邻居节点的计算方式为空间坐标点潜在向量与所述空间中任意一点的点云向量欧氏距离最近的16个节点,欧氏距离:Input the spatial coordinate point potential vector into the voxel decoder to find the set of neighbor node potential vectors of any point in the space, where the neighbor nodes are calculated as the 16 nodes with the closest Euclidean distance between the spatial coordinate point potential vector and the point cloud vector of any point in the space. The Euclidean distance is:
其中xp指邻居节点的三维坐标,yp指邻居节点的三维坐标,zp指邻居节点的三维坐标,xq指空间中任意一点的三维坐标,yq指空间中任意一点的三维坐标,zq指空间中任意一点的三维坐标;Where xp refers to the three-dimensional coordinates of the neighbor node, yp refers to the three-dimensional coordinates of the neighbor node, zp refers to the three-dimensional coordinates of the neighbor node, xq refers to the three-dimensional coordinates of any point in space, yq refers to the three-dimensional coordinates of any point in space, and zq refers to the three-dimensional coordinates of any point in space;
使用插值的方式得到所述空间中任意一点的潜在向量:Use interpolation to get the potential vector of any point in the space:
其中Zq指空间中任意一点的潜在向量,pi∈Nq指邻居节点的潜在向量集合,si∈{Sq}指距离权重;Where Z q refers to the potential vector of any point in space, p i ∈ N q refers to the set of potential vectors of neighboring nodes, and s i ∈ {S q } refers to the distance weight;
空间中任意一点的潜在向量经过一个全连接层后,得到素体点特征向量,多个素体点特征向量构成素体模型;After the potential vector of any point in the space passes through a fully connected layer, a feature vector of the body point is obtained, and multiple feature vectors of the body point constitute a body model;
空间中任意一点的潜在向量经过一个全连接层后,得到空间中任意一点的UV坐标特征向量。After the potential vector of any point in the space passes through a fully connected layer, the UV coordinate feature vector of any point in the space is obtained.
优选地,将所述纹理特征向量和所述UV坐标特征向量输入特征融合模块,得到UV贴图特征向量,包括:Preferably, the texture feature vector and the UV coordinate feature vector are input into a feature fusion module to obtain a UV mapping feature vector, including:
纹理特征向量和UV坐标特征向量相加后输入卷积网络,卷积网络backbone使用ResNet34,得到UV贴图特征向量。The texture feature vector and the UV coordinate feature vector are added and input into the convolutional network. The convolutional network backbone uses ResNet34 to obtain the UV map feature vector.
优选地,在获取家具稀疏点云之后,还包括增强家具稀疏点云数据:Preferably, after obtaining the furniture sparse point cloud, the method further includes enhancing the furniture sparse point cloud data:
将所述家具稀疏点云做高频变换:Perform high-frequency transformation on the furniture sparse point cloud:
其中,为一个家具稀疏点云,/>为实数集,L为高频变度。in, For a furniture sparse point cloud, /> is a set of real numbers, and L is a high frequency variability.
优选地,将所述纹理特征向量和所述UV坐标特征向量输入特征融合模块,得到UV贴图特征向量之后,还包括计算并检验UV贴图和纹理图片之间的相似度,具体为:Preferably, after the texture feature vector and the UV coordinate feature vector are input into a feature fusion module to obtain the UV map feature vector, the method further includes calculating and checking the similarity between the UV map and the texture image, specifically:
将UV贴图特征向量输入torchvision拓展库中,使用transform_convert()方法,输出UV贴图结果,使用基于Gram矩阵的均方误差MSE作为损失函数:Input the UV mapping feature vector into the torchvision extension library, use the transform_convert() method to output the UV mapping result, and use the mean square error MSE based on the Gram matrix as the loss function:
其中C指特征图的通道数,H指特征图的高度,W指特征图的宽度,Gi(Gen)指UV贴图第i层特征图上的Gram矩阵,Gi(Style)指纹理图片在第i层特征图上的Gram矩阵,其中其中Ci为经过神经网络第i层的特征图,/>为Ci的转置矩阵。Where C refers to the number of channels of the feature map, H refers to the height of the feature map, W refers to the width of the feature map, Gi (Gen) refers to the Gram matrix on the feature map of the i-th layer of the UV map, and Gi (Style) refers to the Gram matrix of the texture image on the i-th layer of the feature map. Where Ci is the feature map after the i-th layer of the neural network,/> is the transposed matrix of Ci .
一种基于多模态的素体与纹理融合家具模型重建系统,包括:A furniture model reconstruction system based on multi-modal body and texture fusion, comprising:
数据获取模块,用于获取家具稀疏点云和纹理嵌入向量;Data acquisition module, used to obtain furniture sparse point cloud and texture embedding vector;
数据处理模块,用于将所述家具稀疏点云和所述纹理嵌入向量输入素体与纹理融合模型,得到新纹理家具模型。The data processing module is used to input the furniture sparse point cloud and the texture embedding vector into the body and texture fusion model to obtain a new texture furniture model.
本发明使用基于多模态的体素与纹理融合家具重建方法,家具稀疏点云输入素体编码器和素体解码器,得到素体点特征向量和UV坐标特征向量,纹理图片经过预处理输入纹理编码器,得到纹理特征向量,将UV坐标特征向量和纹理特征向量经过特征融合,得到UV贴图和UV贴图特征向量,素体点特征向量经过open3d拓展库得到素体模型,素体点特征向量与UV贴图特征向量经过特征融合,得到新纹理家具模型。本方法可以使用户在不需要专业家具设计人员的情况下,完成家具模型的纹理更新,大大节省了时间,并且原始数据只用提供二维图片数据,用户可以使用普通相机获得。The present invention uses a furniture reconstruction method based on multi-modal voxel and texture fusion, the furniture sparse point cloud is input into a voxel encoder and a voxel decoder to obtain a voxel point feature vector and a UV coordinate feature vector, the texture image is pre-processed and input into a texture encoder to obtain a texture feature vector, the UV coordinate feature vector and the texture feature vector are subjected to feature fusion to obtain a UV map and a UV map feature vector, the voxel point feature vector is subjected to an open3d expansion library to obtain a voxel model, the voxel point feature vector and the UV map feature vector are subjected to feature fusion to obtain a new texture furniture model. The method allows users to complete the texture update of furniture models without the need for professional furniture designers, greatly saving time, and the original data only needs to provide two-dimensional picture data, which users can obtain using ordinary cameras.
应当理解的是,以上的一般描述和后文的细节描述仅是示例性和解释性的,而非限制本公开。It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the present disclosure.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
为了更清楚地说明本发明实施例或背景技术中的技术方案,下面将对本发明实施例或背景技术中所需要使用的附图进行说明。In order to more clearly illustrate the technical solutions in the embodiments of the present invention or the background technology, the drawings required for use in the embodiments of the present invention or the background technology will be described below.
此处的附图被并入说明书中并构成本说明书的一部分,这些附图示出了符合本公开的实施例,并与说明书一起用于说明本发明公开的技术方案。The drawings herein are incorporated into the specification and constitute a part of the specification. These drawings illustrate embodiments consistent with the present disclosure and, together with the specification, are used to illustrate the technical solutions disclosed in the present invention.
图1为本发明实例中的基于多模态的素体与纹理融合家具模型重建方法流程图;FIG1 is a flow chart of a method for reconstructing a furniture model based on multi-modal body and texture fusion in an example of the present invention;
图2为本发明实例中获取家具稀疏点云的流程图;FIG2 is a flow chart of obtaining a sparse point cloud of furniture in an example of the present invention;
图3为本发明实例中素体与纹理融合模块结构图;FIG3 is a structural diagram of a body and texture fusion module in an example of the present invention;
图4为本发明实例中的素体编码器结构图;FIG4 is a structural diagram of a body encoder in an example of the present invention;
图5为本发明实例中的素体解码器结构图;FIG5 is a structural diagram of a body decoder in an example of the present invention;
图6为本发明实例中的纹理编码器结构图;FIG6 is a structural diagram of a texture encoder in an example of the present invention;
图7为本发明实例中的特征融合模块结构图;FIG7 is a structural diagram of a feature fusion module in an example of the present invention;
图8为本发明实例中的素体重建效果图;FIG8 is a diagram showing the reconstruction effect of a body in an example of the present invention;
图9为本发明实例中的纹理更新家具模型效果图。FIG. 9 is a rendering of a furniture model with texture update in an example of the present invention.
具体实施方式Detailed ways
为了使本技术领域的人员更好地理解本发明方案,下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。In order to enable those skilled in the art to better understand the scheme of the present invention, the technical scheme in the embodiments of the present invention will be clearly and completely described below in conjunction with the drawings in the embodiments of the present invention. Obviously, the described embodiments are only part of the embodiments of the present invention, not all of the embodiments. Based on the embodiments of the present invention, all other embodiments obtained by ordinary technicians in this field without creative work are within the scope of protection of the present invention.
本发明的说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别不同对象,而不是用于描述特定顺序。此外,术语“包括”和“具有”以及它们任何变形,意图在于覆盖不排他的包含。例如包含了一系列步骤或单元的过程、方法、系统、产品或设备没有限定于已列出的步骤或单元,而是可选地还包括没有列出的步骤或单元,或可选地还包括对于这些过程、方法、产品或设备固有的其他步骤或单元。The terms "first", "second", etc. in the specification and claims of the present invention and the above-mentioned drawings are used to distinguish different objects, rather than to describe a specific order. In addition, the terms "including" and "having" and any variations thereof are intended to cover non-exclusive inclusions. For example, a process, method, system, product or device that includes a series of steps or units is not limited to the listed steps or units, but may optionally include steps or units that are not listed, or may optionally include other steps or units that are inherent to these processes, methods, products or devices.
本文中术语“和/或”,仅仅是一种描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。另外,本文中术语“至少一种”表示多种中的任意一种或多种中的至少两种的任意组合,例如,包括A、B、C中的至少一种,可以表示包括从A、B和C构成的集合中选择的任意一个或多个元素。The term "and/or" herein is only a description of the association relationship of the associated objects, indicating that there may be three relationships. For example, A and/or B can represent: A exists alone, A and B exist at the same time, and B exists alone. In addition, the term "at least one" herein represents any combination of at least two of any one or more of a plurality of. For example, including at least one of A, B, and C can represent including any one or more elements selected from the set consisting of A, B, and C.
在本文中提及“实施例”意味着,结合实施例描述的特定特征、结构或特性可以包含在本发明的至少一个实施例中。在说明书中的各个位置出现该短语并不一定均是指相同的实施例,也不是与其它实施例互斥的独立的或备选的实施例。本领域技术人员显式地和隐式地理解的是,本文所描述的实施例可以与其它实施例相结合。Reference to "embodiments" herein means that a particular feature, structure, or characteristic described in conjunction with the embodiments may be included in at least one embodiment of the present invention. The appearance of the phrase in various places in the specification does not necessarily refer to the same embodiment, nor is it an independent or alternative embodiment that is mutually exclusive with other embodiments. It is explicitly and implicitly understood by those skilled in the art that the embodiments described herein may be combined with other embodiments.
另外,为了更好地说明本发明,在下文的具体实施方式中给出了众多的具体细节。本领域技术人员应当理解,没有某些具体细节,本发明同样能够实施。在一些实例中,对于本领域技术人员熟知的方法、手段、元件和电路未作详细描述,以便于凸显本发明的主旨。In addition, in order to better illustrate the present invention, numerous specific details are provided in the following specific embodiments. It should be understood by those skilled in the art that the present invention can be implemented without certain specific details. In some examples, methods, means, components and circuits well known to those skilled in the art are not described in detail in order to highlight the subject matter of the present invention.
随着家具的发展与互联网的普及,家具商可以通过网络媒介展示各种家具效果,不再需要客户亲自前往家具市场查看家具实物,大大节省了时间,但是制作一个可供客户在网络上参照的家具模型,需要专业的家具设计人员或建模工程师等花费大量时间进行3D建模,商家需要一种新的方法来提高家具建模的效率。With the development of furniture and the popularization of the Internet, furniture dealers can display various furniture effects through online media. Customers no longer need to go to the furniture market in person to view the actual furniture, which greatly saves time. However, to make a furniture model for customers to refer to on the Internet, professional furniture designers or modeling engineers need to spend a lot of time on 3D modeling. Merchants need a new method to improve the efficiency of furniture modeling.
使用传统计算机视觉的方法中,通常使用多个2D图像或视频来构建3D模型。例如使用多视角几何的方法,将多个图像或视频帧中的像素坐标转换为3D点,然后使用这些点云数据生成家具模型,从每张图像中选择重要特征是必要步骤,而且随着类别数量的增加,特征提取会得越来越麻烦。使用深度学习的计算机视觉方法中,通常用神经网络模型来从2D图像或点云数据中预测家具的三维模型。例如,NeRF神经网络体渲染技术,能达到不错的效果,但该模型训练时间太长,且缺乏泛化能力。Instant-NGP创造性地引入hash编码到神经网络中,提升了模型推理的速度,但其导出的三维模型表面纹理效果不佳和模型消耗显存过大,使得在家具领域仍旧没有很好的应用。现如今,算力的提升和数据的丰富使得深度学习发展迅速,其中多模态神经网络是指能够处理多个输入数据类型的神经网络,本发明使用家具稀疏点云和纹理图片作为输入,经过神经网络,预测出素体模型,UV贴图以及新纹理家具模型,能够大大节省家具设计人员制作家具模型的时间,大大节省了时间,并且原始数据只用提供二维图片数据,用户可以使用普通相机获得。In traditional computer vision methods, multiple 2D images or videos are usually used to build a 3D model. For example, using multi-view geometry methods, pixel coordinates in multiple images or video frames are converted into 3D points, and then these point cloud data are used to generate furniture models. Selecting important features from each image is a necessary step, and as the number of categories increases, feature extraction becomes more and more troublesome. In computer vision methods using deep learning, neural network models are usually used to predict the 3D model of furniture from 2D images or point cloud data. For example, NeRF neural network volume rendering technology can achieve good results, but the model takes too long to train and lacks generalization ability. Instant-NGP creatively introduces hash coding into the neural network, which improves the speed of model reasoning, but the surface texture of the exported 3D model is not good and the model consumes too much video memory, so it is still not well applied in the furniture field. Nowadays, the improvement of computing power and the abundance of data have led to the rapid development of deep learning. Among them, multimodal neural networks refer to neural networks that can process multiple input data types. The present invention uses furniture sparse point clouds and texture images as input, and predicts the body model, UV map and new texture furniture model through the neural network, which can greatly save the time of furniture designers in making furniture models, greatly saving time, and the original data only needs to provide two-dimensional image data, which users can obtain using ordinary cameras.
本发明使用家具稀疏点云和纹理图片作为输入,经过神经网络,预测出素体模型,UV贴图以及新纹理家具模型,能够大大节省家具设计人员制作家具模型的时间,大大节省了时间,并且原始数据只用提供二维图片数据,用户可以使用普通相机获得。The present invention uses furniture sparse point cloud and texture pictures as input, and predicts the body model, UV map and new texture furniture model through a neural network, which can greatly save the time of furniture designers in making furniture models, greatly saving time, and the original data only needs to provide two-dimensional picture data, which users can obtain using an ordinary camera.
实施例1Example 1
一种基于多模态的素体与纹理融合家具模型重建方法,参考图1,包括:A furniture model reconstruction method based on multi-modal body and texture fusion, referring to FIG1, comprises:
S100获取家具稀疏点云和纹理嵌入向量,如图2,拍摄原始家具图片不少于20张,作为家具图片集,确保相邻两张照片移动角度不超过45°,否则无法构建位姿信息,也无法输出家具稀疏点云,使用SFM软件colmap处理得到家具稀疏点云;S100 obtains furniture sparse point cloud and texture embedding vector, as shown in Figure 2. Take at least 20 original furniture pictures as the furniture picture set, and ensure that the shift angle between two adjacent pictures does not exceed 45°, otherwise the pose information cannot be constructed, and the furniture sparse point cloud cannot be output. Use SFM software colmap to process and obtain the furniture sparse point cloud;
在colmap中新建工程项目,创建项目数据库,导入家具图片集;Create a new project in colmap, create a project database, and import the furniture picture collection;
使用feature extraction功能进行特征提取,使用feature match功能进行特征匹配,生成场景图和匹配矩阵;Use the feature extraction function to extract features, use the feature match function to match features, and generate scene graphs and matching matrices;
使用start reconstruction功能进行增量式重建,colmap图形用户界面将生成场景的家具稀疏点云和各个视角的相机姿态,获得家具稀疏点云重建结果,并设置点云信息只包含空间坐标与灰度值;Use the start reconstruction function to perform incremental reconstruction. The colmap graphical user interface will generate the furniture sparse point cloud of the scene and the camera poses of each viewpoint, obtain the furniture sparse point cloud reconstruction result, and set the point cloud information to only contain spatial coordinates and grayscale values;
将家具稀疏点云重建结果保存为文本,得到家具稀疏点云,每个点云数据的维度为4维,分别为(x,y,z,g);其中x,y,z为世界坐标系下点坐标,g为点云灰度值;The furniture sparse point cloud reconstruction result is saved as text to obtain the furniture sparse point cloud. The dimension of each point cloud data is 4 dimensions, namely (x, y, z, g); x, y, z are the point coordinates in the world coordinate system, and g is the gray value of the point cloud;
选择一张需要附着在家具表面的纹理图片,将纹理图片缩放为长224像素,宽224像素,设置为RGB3通道,得到标准纹理图片;Select a texture image that needs to be attached to the surface of the furniture, scale the texture image to 224 pixels in length and 224 pixels in width, set it to RGB3 channels, and get a standard texture image;
将标准纹理图片切分为长16像素,宽16像素的碎片patch,一张标准纹理图片可以得到196个patch,每个patch维度为768;The standard texture image is divided into patches with a length of 16 pixels and a width of 16 pixels. A standard texture image can get 196 patches, and the dimension of each patch is 768;
将得到的196个patch输入全连接层,全连接层的输入节点是768,输出节点是768,输出[196,768]的无位置信息的patch向量;The obtained 196 patches are input into the fully connected layer. The input node of the fully connected layer is 768, the output node is 768, and the output is a patch vector [196, 768] without position information.
将无位置信息的patch向量加上一维位置编码,一维位置编码与无位置信息的patch向量直接相加,Epixel为一维位置编码,得到纹理嵌入向量Epatch;The patch vector without position information is added with a one-dimensional position code. The one-dimensional position code is directly added to the patch vector without position information. E pixel is the one-dimensional position code. Get the texture embedding vector E patch ;
S200将家具稀疏点云和纹理嵌入向量输入素体与纹理融合模型,得到新纹理家具模型,如图9。S200 inputs the furniture sparse point cloud and texture embedding vector into the body and texture fusion model to obtain a new texture furniture model, as shown in Figure 9.
优选地,所述素体与纹理融合模型,包括:Preferably, the body and texture fusion model includes:
素体编码器,用于根据家具稀疏点云生成空间坐标点潜在向量;The body encoder is used to generate the spatial coordinate point potential vector based on the furniture sparse point cloud;
素体解码器,用于根据空间坐标点潜在向量生成素体点特征向量和UV坐标特征向量;A voxel decoder is used to generate a voxel point feature vector and a UV coordinate feature vector according to a spatial coordinate point potential vector;
纹理编码器,用于根据纹理嵌入向量生成纹理特征向量;A texture encoder, for generating a texture feature vector according to the texture embedding vector;
特征融合模块,用于根据纹理特征向量和UV坐标特征向量生成UV贴图特征向量,并根据UV贴图特征向量和素体点特征向量生成新纹理家具模型;A feature fusion module is used to generate a UV mapping feature vector according to a texture feature vector and a UV coordinate feature vector, and to generate a new texture furniture model according to the UV mapping feature vector and a body point feature vector;
基于全连接层的素体编码器,用于根据家具稀疏点云生成空间坐标点潜在向量;A fully connected layer-based body encoder is used to generate spatial coordinate point latent vectors based on the furniture sparse point cloud;
基于插值的素体解码器,用于根据空间坐标点潜在向量生成素体点特征向量和UV坐标特征向量;Interpolation-based voxel decoder, used to generate voxel point feature vectors and UV coordinate feature vectors based on spatial coordinate point latent vectors;
基于多头注意力的纹理编码器,用于根据纹理嵌入向量生成纹理特征向量;A multi-head attention-based texture encoder is used to generate a texture feature vector from a texture embedding vector;
基于卷积神经网络的特征融合模块,如图7,用于根据纹理特征向量和UV坐标特征向量生成UV贴图特征向量,并根据UV贴图特征向量和素体点特征向量生成新纹理家具模型。The feature fusion module based on the convolutional neural network, as shown in FIG7 , is used to generate a UV mapping feature vector according to the texture feature vector and the UV coordinate feature vector, and to generate a new texture furniture model according to the UV mapping feature vector and the body point feature vector.
优选地,S200将家具稀疏点云和纹理嵌入向量输入素体与纹理融合模型,得到新纹理家具模型,参考图3,包括:Preferably, S200 inputs the furniture sparse point cloud and texture embedding vector into the body and texture fusion model to obtain a new texture furniture model, referring to FIG3 , including:
S210将家具稀疏点云输入素体编码器,如图4,,得到空间坐标点潜在向量:S210 inputs the furniture sparse point cloud into the body encoder, as shown in Figure 4, and obtains the potential vector of the spatial coordinate point:
增强家具稀疏点云数据输入素体编码器,素体编码器为8层的全连接层网络,第1~7层的神经元个数为256,第8层的神经元个数为125,在第1全连接层和第4全连接层做残差连接,最终网络输出维度大小为125的潜在向量;The sparse point cloud data of the enhanced furniture is input into the body encoder. The body encoder is an 8-layer fully connected network. The number of neurons in the 1st to 7th layers is 256, and the number of neurons in the 8th layer is 125. Residual connections are made in the 1st and 4th fully connected layers. Finally, the network outputs a potential vector with a dimension size of 125.
每个家具稀疏点云的空间坐标和对应潜在向量做一次拼接操作,维度升至128维,得到空间坐标点潜在向量;The spatial coordinates of each furniture sparse point cloud and the corresponding latent vector are concatenated once, and the dimension is increased to 128 dimensions to obtain the latent vector of the spatial coordinate point;
S220将空间坐标点潜在向量输入素体解码器,得到素体点特征向量和UV坐标特征向量;S220 inputs the spatial coordinate point potential vector into the body decoder to obtain the body point feature vector and the UV coordinate feature vector;
S230将纹理嵌入向量输入纹理编码器,得到纹理特征向量;S230 inputs the texture embedding vector into a texture encoder to obtain a texture feature vector;
S240将纹理特征向量和UV坐标特征向量输入特征融合模块,得到UV贴图特征向量;S240 inputs the texture feature vector and the UV coordinate feature vector into a feature fusion module to obtain a UV mapping feature vector;
S250将UV贴图特征向量与素体点特征向量相加,经过卷积网络,使用open3D拓展库,得到新纹理家具模型;S250 adds the UV map feature vector to the body point feature vector, passes through the convolutional network, and uses the open3D extension library to obtain a new texture furniture model;
UV贴图特征向量Tq和素体特征向量Rq经过卷积网络,卷积网络backbone使用ResNet34,通过open3d拓展库,输出新纹理家具模型。The UV map feature vector T q and the body feature vector R q pass through the convolutional network. The convolutional network backbone uses ResNet34 and outputs a new texture furniture model through the open3d extension library.
优选地,S230纹理嵌入向量输入纹理编码器,输出纹理特征向量,参考图6,包括:Preferably, S230 texture embedding vector is input into a texture encoder, and a texture feature vector is output, referring to FIG6 , including:
将有位置信息的纹理嵌入向量Epatch,经过层归一化后将层归一化结果和多头注意力计算结果残差连接,得到纹理潜在向量E‘patch:The texture with position information is embedded in the vector E patch . After layer normalization, the layer normalization result and the multi-head attention calculation result are residually connected to obtain the texture potential vector E' patch :
E‘patch=MSA(LN(Epatch)+Epatch E'patch =MSA(LN( Epatch )+ Epatch
其中E‘patch是纹理潜在向量,Epatch是有位置信息的纹理嵌入向量;Where E'patch is the texture potential vector, and Epatch is the texture embedding vector with position information;
纹理潜在向量E‘patch经过层归一化,全连接层,后将层归一化和全连接层计算结果做残差计算,输出纹理特征向量T:The texture potential vector E'patch is normalized and fully connected. The residual calculation is performed on the layer normalization and fully connected layer calculation results to output the texture feature vector T:
T=MLP(LN(E‘patch))+E‘patch T=MLP(LN( E'patch ))+ E'patch
其中T是纹理特征向量,E‘patch是纹理潜在向量。Where T is the texture feature vector and E'patch is the texture latent vector.
优选地,S220将空间坐标点潜在向量输入素体解码器,得到素体点特征向量和UV坐标特征向量,参考图5,包括:Preferably, S220 inputs the spatial coordinate point potential vector into the body decoder to obtain the body point feature vector and the UV coordinate feature vector, referring to FIG5 , including:
空间坐标点潜在向量输入素体解码器,需要构造空间中除家具稀疏点云外任意一点q的点云向量Zq,Zq是当前只包含空间坐标的128维向量,前3维为空间坐标,后面全部设置为0;The spatial coordinate point potential vector is input into the volume decoder, which needs to construct the point cloud vector Z q of any point q in the space except the furniture sparse point cloud. Z q is a 128-dimensional vector that currently only contains spatial coordinates. The first 3 dimensions are spatial coordinates, and the rest are all set to 0.
将空间坐标点潜在向量输入素体解码器,找到空间中任意一点q的邻居节点潜在向量集合pi∈{Nq},其中邻居节点的计算方式为空间坐标点潜在向量与空间中任意一点q的点云向量Zq欧氏距离最近的16个节点,欧氏距离:Input the spatial coordinate point potential vector into the voxel decoder and find the neighbor node potential vector set p i ∈ {N q } of any point q in space. The neighbor nodes are calculated as the 16 nodes with the closest Euclidean distance between the spatial coordinate point potential vector and the point cloud vector Z q of any point q in space. The Euclidean distance is:
其中xp指邻居节点的三维坐标,yp指邻居节点的三维坐标,zp指邻居节点的三维坐标,xq指空间中任意一点的三维坐标,yq指空间中任意一点的三维坐标,zq指空间中任意一点的三维坐标;Where xp refers to the three-dimensional coordinates of the neighbor node, yp refers to the three-dimensional coordinates of the neighbor node, zp refers to the three-dimensional coordinates of the neighbor node, xq refers to the three-dimensional coordinates of any point in space, yq refers to the three-dimensional coordinates of any point in space, and zq refers to the three-dimensional coordinates of any point in space;
使用插值的方式得到空间中任意一点的潜在向量:Use interpolation to get the potential vector at any point in space:
其中Zq指空间中任意一点的潜在向量,pi∈Nq指邻居节点的潜在向量集合,si∈{Sq}指距离权重;Where Z q refers to the potential vector of any point in space, p i ∈ N q refers to the set of potential vectors of neighboring nodes, and s i ∈ {S q } refers to the distance weight;
式中si:⊙为hardmard product为可学习的距离权重,对于查询点q来说,离其距离越近的邻居节点,所用权重||si||2范数更大,代表该邻居节点的潜在向量对查询点q关联系更强;Where s i : ⊙ is the hardmard product, which is a learnable distance weight. For a query point q, the closer the neighbor node is to it, the larger the norm of the weight ||s i || 2 is, which means that the potential vector of the neighbor node is more closely related to the query point q.
空间中任意一点q的潜在向量Zq经过一个全连接层后,得到素体点特征向量Rq,多个素体点特征向量Rq构成素体模型;After the potential vector Z q of any point q in the space passes through a fully connected layer, the feature vector R q of the body point is obtained. Multiple feature vectors R q of the body point constitute the body model.
空间中任意一点q的潜在向量Zq经过一个全连接层后,得到空间中任意一点的UV坐标特征向量Cq。After the potential vector Z q of any point q in the space passes through a fully connected layer, the UV coordinate feature vector C q of any point in the space is obtained.
优选地,S240将纹理特征向量和UV坐标特征向量输入特征融合模块,得到UV贴图特征向量,如图7,包括:Preferably, S240 inputs the texture feature vector and the UV coordinate feature vector into a feature fusion module to obtain a UV mapping feature vector, as shown in FIG7 , including:
纹理特征向量T和UV坐标特征向量Cq相加后输入卷积网络,卷积网络backbone使用ResNet34,得到UV贴图特征向量Tq;The texture feature vector T and the UV coordinate feature vector C q are added and input into the convolutional network. The convolutional network backbone uses ResNet34 to obtain the UV map feature vector T q ;
UV贴图特征向量Tq和素体特征向量Rq经过卷积网络,卷积网络backbone使用ResNet34,通过open3d拓展库,输出新纹理家具模型。The UV map feature vector T q and the body feature vector R q pass through the convolutional network. The convolutional network backbone uses ResNet34 and outputs a new texture furniture model through the open3d extension library.
优选地,多个素体点特征向量通过open3d拓展库构成素体模型包括:Preferably, the plurality of element point feature vectors constitute an element model through an open3d extension library, including:
使用Chamfer Loss作为损失函数生成素体模型,如图8,具体为:Use Chamfer Loss as the loss function to generate the body model, as shown in Figure 8, specifically:
其中pi是素体点云中的第i个点,qj是真实点云中的第j个点,Np是素体点云中的数量,Ngt是真实点云中的数量。Where pi is the i-th point in the body point cloud, qj is the j-th point in the real point cloud, Np is the number in the body point cloud, and Ngt is the number in the real point cloud.
优选地,S200将家具稀疏点云和纹理嵌入向量输入素体与纹理融合模型,得到新纹理家具模型之前,还包括对纹理与素体融合模型进行训练,具体为:Preferably, S200 inputs the furniture sparse point cloud and texture embedding vector into the body and texture fusion model to obtain a new texture furniture model, and further includes training the texture and body fusion model, specifically:
通过IKEA 3D Models数据集获取家具稀疏点云,使用CUReT数据集作为纹理图片获取纹理嵌入向量;The IKEA 3D Models dataset is used to obtain furniture sparse point clouds, and the CUReT dataset is used as a texture image to obtain texture embedding vectors.
将IKEA 3D Models数据集家具稀疏点云和CUReT数据集纹理图片纹理嵌入向量输入素体与纹理融合模型;The IKEA 3D Models dataset furniture sparse point cloud and the CUReT dataset texture image texture embedding vectors are input into the volume and texture fusion model;
素体与纹理融合模型输出新纹理家具模型;The body and texture fusion model outputs a new texture furniture model;
迭代训练至模型收敛。Iterate training until the model converges.
优选地,S100获取家具稀疏点云之后,还包括S110增强稀疏点云数据:Preferably, after obtaining the furniture sparse point cloud in S100, the method further includes enhancing the sparse point cloud data in S110:
将家具稀疏点云做高频变换:Perform high-frequency transformation on the furniture sparse point cloud:
其中,为一个家具稀疏点云,/>为实数集,L为高频变度;in, For a furniture sparse point cloud, /> is a set of real numbers, L is a high frequency variability;
高频变度L设置为3,点云数据维度从4增至24,得到增强家具稀疏点云数据。The high-frequency variation L is set to 3, and the point cloud data dimension is increased from 4 to 24, obtaining enhanced furniture sparse point cloud data.
优选地,S240将纹理特征向量和UV坐标特征向量输入特征融合模块,得到UV贴图特征向量之后,还包括计算并检验UV贴图和纹理图片之间的相似度,具体为:Preferably, after inputting the texture feature vector and the UV coordinate feature vector into the feature fusion module to obtain the UV map feature vector, S240 also includes calculating and checking the similarity between the UV map and the texture image, specifically:
将UV贴图特征向量Tq输入torchvision拓展库中,使用transform_convert()方法,输出UV贴图结果,UV贴图本质上为纹理图片的风格迁移,所以要计算UV贴图和纹理图片之间的相似度,使用基于Gram矩阵的均方误差MSE作为损失函数:Input the UV map feature vector T q into the torchvision extension library, use the transform_convert() method to output the UV map result. The UV map is essentially the style transfer of the texture image, so to calculate the similarity between the UV map and the texture image, use the mean square error MSE based on the Gram matrix as the loss function:
其中C指特征图的通道数,H指特征图的高度,W指特征图的宽度,Gi(Gen)指UV贴图第i层特征图上的Gram矩阵,Gi(Style)指纹理图片在第i层特征图上的Gram矩阵,其中其中Ci为经过神经网络第i层的特征图,/>为Ci的转置矩阵。Where C refers to the number of channels of the feature map, H refers to the height of the feature map, W refers to the width of the feature map, Gi (Gen) refers to the Gram matrix on the feature map of the i-th layer of the UV map, and Gi (Style) refers to the Gram matrix of the texture image on the i-th layer of the feature map. Where Ci is the feature map after the i-th layer of the neural network,/> is the transposed matrix of Ci .
实施例2Example 2
一种基于多模态的素体与纹理融合家具模型重建系统,包括:A furniture model reconstruction system based on multi-modal body and texture fusion, comprising:
数据获取模块,用于获取家具稀疏点云和纹理嵌入向量;Data acquisition module, used to obtain furniture sparse point cloud and texture embedding vector;
数据处理模块,用于将家具稀疏点云和纹理嵌入向量输入素体与纹理融合模型,得到新纹理家具模型。The data processing module is used to input the furniture sparse point cloud and texture embedding vector into the body and texture fusion model to obtain a new texture furniture model.
Claims (6)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202310714863.4A CN116740300B (en) | 2023-06-16 | 2023-06-16 | Multi-mode-based prime body and texture fusion furniture model reconstruction method |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202310714863.4A CN116740300B (en) | 2023-06-16 | 2023-06-16 | Multi-mode-based prime body and texture fusion furniture model reconstruction method |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN116740300A CN116740300A (en) | 2023-09-12 |
| CN116740300B true CN116740300B (en) | 2024-05-03 |
Family
ID=87911097
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202310714863.4A Active CN116740300B (en) | 2023-06-16 | 2023-06-16 | Multi-mode-based prime body and texture fusion furniture model reconstruction method |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN116740300B (en) |
Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN103150751A (en) * | 2013-01-10 | 2013-06-12 | 江苏易图地理信息工程有限公司 | Three-dimensional modeling method for achieving building inside and outside integration in digital map |
| CN109978930A (en) * | 2019-03-27 | 2019-07-05 | 杭州相芯科技有限公司 | A kind of stylized human face three-dimensional model automatic generation method based on single image |
| CN115272636A (en) * | 2022-07-28 | 2022-11-01 | 北京优酷科技有限公司 | Method and device for generating digital collection model and electronic equipment |
| CN115423978A (en) * | 2022-08-30 | 2022-12-02 | 西北工业大学 | Deep Learning-Based Image Laser Data Fusion Method for Building Reconstruction |
| CN115908517A (en) * | 2023-01-06 | 2023-04-04 | 广东工业大学 | Low-overlap point cloud registration method based on corresponding point matching matrix optimization |
| CN116152267A (en) * | 2023-04-24 | 2023-05-23 | 中国民用航空飞行学院 | Point Cloud Instance Segmentation Method Based on Contrastive Language Image Pre-training Technology |
Family Cites Families (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| RU2215326C2 (en) * | 2001-06-29 | 2003-10-27 | Самсунг Электроникс Ко., Лтд. | Image-based hierarchic presentation of motionless and animated three-dimensional object, method and device for using this presentation to visualize the object |
-
2023
- 2023-06-16 CN CN202310714863.4A patent/CN116740300B/en active Active
Patent Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN103150751A (en) * | 2013-01-10 | 2013-06-12 | 江苏易图地理信息工程有限公司 | Three-dimensional modeling method for achieving building inside and outside integration in digital map |
| CN109978930A (en) * | 2019-03-27 | 2019-07-05 | 杭州相芯科技有限公司 | A kind of stylized human face three-dimensional model automatic generation method based on single image |
| CN115272636A (en) * | 2022-07-28 | 2022-11-01 | 北京优酷科技有限公司 | Method and device for generating digital collection model and electronic equipment |
| CN115423978A (en) * | 2022-08-30 | 2022-12-02 | 西北工业大学 | Deep Learning-Based Image Laser Data Fusion Method for Building Reconstruction |
| CN115908517A (en) * | 2023-01-06 | 2023-04-04 | 广东工业大学 | Low-overlap point cloud registration method based on corresponding point matching matrix optimization |
| CN116152267A (en) * | 2023-04-24 | 2023-05-23 | 中国民用航空飞行学院 | Point Cloud Instance Segmentation Method Based on Contrastive Language Image Pre-training Technology |
Non-Patent Citations (1)
| Title |
|---|
| 适老家具的CMF设计研究;崔敦睿 等;《林产工业》;20200506(第05期);第61-64页 * |
Also Published As
| Publication number | Publication date |
|---|---|
| CN116740300A (en) | 2023-09-12 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US20220058429A1 (en) | Method for fine-grained sketch-based scene image retrieval | |
| CN116310076A (en) | Three-dimensional reconstruction method, device, equipment and storage medium based on nerve radiation field | |
| WO2021143264A1 (en) | Image processing method and apparatus, server and storage medium | |
| CN112465122B (en) | Device and method for optimizing original dimension operator in neural network model | |
| CN113850900A (en) | Method and system for recovering depth map based on image and geometric clue in three-dimensional reconstruction | |
| CN114758205B (en) | Multi-view feature fusion method and system for 3D human posture estimation | |
| CN118298109B (en) | View processing method for multimodal electronic information system | |
| Sun et al. | Quadratic terms based point-to-surface 3D representation for deep learning of point cloud | |
| CN118397282B (en) | Three-dimensional point cloud robustness component segmentation method based on semantic SAM large model | |
| CN120107568B (en) | Multi-source target detection method based on pixel-level dynamic mode selection fusion | |
| CN117576312A (en) | Hand model construction method and device and computer equipment | |
| CN116912296A (en) | Point cloud registration method based on position-enhanced attention mechanism | |
| Zhang et al. | Hybrid feature CNN model for point cloud classification and segmentation | |
| CN116188894A (en) | Neural rendering-based point cloud pre-training method, system, device and medium | |
| CN113158970B (en) | Action identification method and system based on fast and slow dual-flow graph convolutional neural network | |
| CN116740300B (en) | Multi-mode-based prime body and texture fusion furniture model reconstruction method | |
| CN114596345A (en) | Point cloud registration method and device, electronic equipment and computer storage medium | |
| CN119580034A (en) | Training method for generating picture description model, picture description generation method, device, equipment, medium and program product | |
| CN116844070B (en) | Point cloud data processing method for 3D modeling of UAV aerial photography based on multi-dimensional shape adaptive convolution | |
| CN118710695A (en) | A method, product, medium and device for reconstructing depth map of RGB-D image | |
| CN118015192A (en) | A fast 3D model reconstruction method based on 2D images | |
| CN117392224A (en) | Zero sample domain migration method and system based on reconstruction-simulation-perception automatic driving | |
| CN114596361B (en) | Depth map and pose optimization method and system | |
| CN113205579B (en) | Three-dimensional reconstruction method, device, equipment and storage medium | |
| CN114373059B (en) | Sketch-based 3D modeling method for smart clothing |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |