[go: up one dir, main page]

CN114048822A - An Image Attention Mechanism Feature Fusion Segmentation Method - Google Patents

An Image Attention Mechanism Feature Fusion Segmentation Method Download PDF

Info

Publication number
CN114048822A
CN114048822A CN202111398830.0A CN202111398830A CN114048822A CN 114048822 A CN114048822 A CN 114048822A CN 202111398830 A CN202111398830 A CN 202111398830A CN 114048822 A CN114048822 A CN 114048822A
Authority
CN
China
Prior art keywords
feature
feature map
output
attention mechanism
fusion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111398830.0A
Other languages
Chinese (zh)
Inventor
刘辉
曲长波
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Liaoning Technical University
Original Assignee
Liaoning Technical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Liaoning Technical University filed Critical Liaoning Technical University
Priority to CN202111398830.0A priority Critical patent/CN114048822A/en
Publication of CN114048822A publication Critical patent/CN114048822A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开了一种图像的注意力机制特征融合分割方法,步骤为:将输入图像统一裁剪;对输出特征图进行池化操作;将特征提取网络ResNet101中四个不同的卷积阶段;将输出特征图通过空间金字塔池化聚合多尺度的上下文信息;对输出特征图通过双线性插值进行两倍上采样,经通道注意力机制与输出特征图进行特征融合;对输出特征图通过双线性插值进行两倍上采样,经通道注意力机制与输出特征图进行特征融合,作为通道注意力机制特征融合模块的输出特征图;使用双线性插值对特诊图进行4倍上采样,还原至输入图像分辨率,生成最终预测结果。本发明有助于形成更加清晰的目标分割边界和精细的分割结果,充分利用低级特征图中的细节信息,提高图像的语义分割精度。

Figure 202111398830

The invention discloses an image attention mechanism feature fusion segmentation method. The steps are as follows: uniformly crop the input image; perform a pooling operation on the output feature map; extract four different convolution stages in the feature extraction network ResNet101; The feature map aggregates multi-scale context information through spatial pyramid pooling; double upsampling the output feature map through bilinear interpolation, and performs feature fusion with the output feature map through the channel attention mechanism; The interpolation is up-sampling twice, and the channel attention mechanism and the output feature map are used for feature fusion, as the output feature map of the channel attention mechanism feature fusion module; bilinear interpolation is used to upsample the special diagnosis map by 4 times, and restore to Enter the image resolution to generate the final prediction. The invention helps to form a clearer target segmentation boundary and a finer segmentation result, makes full use of the detail information in the low-level feature map, and improves the semantic segmentation accuracy of the image.

Figure 202111398830

Description

Attention mechanism feature fusion segmentation method for image
Technical Field
The invention belongs to the technical field of image semantic segmentation, and particularly relates to an attention mechanism feature fusion segmentation method for an image.
Background
With the continuous progress of society, people expect that computers can have logical reasoning ability and decision making ability like human beings, and people are saved from various complex tasks. Image semantic segmentation divides an image into regions representing different semantic information by assigning a semantic label which is well defined in advance to represent a category to a pixel in the image. The image subjected to image semantic segmentation can be used for scene understanding tasks such as image semantic recognition, target tracking and the like, and is an important means for image processing. At present, image semantic segmentation technology has been applied to the fields of medical imaging, automatic driving, intelligent home, image engine searching and the like.
The Image Semantic Segmentation method can be divided into a conventional Image Semantic Segmentation method and an Image Semantic Segmentation method (ISSbDL) Based On Deep Learning, and is hereinafter referred to as a Deep Semantic Segmentation method. Most of the conventional semantic segmentation methods perform segmentation according to the characteristics of the gray scale, texture and the like of an image, such as an image pixel threshold-based segmentation method, an image object edge-based segmentation method, a region-based segmentation method or a super-pixel segmentation method. The traditional semantic segmentation algorithm needs to manually extract features, and the quality of a segmentation result depends on the quality of the feature extraction result, so that the traditional semantic segmentation method is time-consuming and labor-consuming and the segmentation result is rough. The deep semantic segmentation method can automatically extract image features through the powerful feature extraction capability of a Convolutional Neural Network (CNN) so as to carry out end-to-end training, deep Labv3+ adopts ResNet101 as a feature extraction Network, collects and multiscale context information through a spatial pyramid pooling, and finally adopts a simple and efficient decoder to restore a prediction score image to the size of an input image to serve as a prediction result.
In the deep label bv3+ network, only the feature map with the down-sampling step length of 4 is fused in the decoder, shallow feature map information in the convolution process is not fully utilized, and the interested target cannot be subjected to side learning, so that the segmentation effect of the target boundary pixel points is rough, and the small target is easy to miss segmentation.
Image semantic segmentation is one of key problems in computer vision, and a convolutional neural network is a mainstream method of a semantic segmentation task. The encoder of the DeepLabv3+ semantic segmentation network effectively extracts high-level features, but a decoder directly fuses a single low-level feature map and a high-level feature map in the feature extraction network, the feature fusion mode is too simple, and the detail information of an image cannot be effectively recovered, so that the target edge pixels in a segmentation result are not accurately positioned, and the problems of missing segmentation and wrong segmentation exist.
Disclosure of Invention
Based on the defects of the prior art, the technical problem to be solved by the invention is to provide an attention mechanism feature fusion segmentation method for an image, which is helpful for forming clearer target segmentation boundaries and fine segmentation results, and fully utilizes detail information in a low-level feature map to improve the semantic segmentation precision of the image.
In order to realize the invention, the invention provides an attention mechanism feature fusion segmentation method of an image, which comprises the following steps:
s1, uniformly cutting the input image into 513 × 513 resolution, and cutting the input image with the original size smaller than 513 × 513 after zero filling operation;
s2, reducing the input image size from 513 × 513 to 257 × 257 by 7 × 7 convolution with convolution step size 2;
s3, performing pooling operation with step size of 2 and pooling kernel size of 3 × 3 on the output feature map of step S2, and reducing the feature map size to 129 × 129 by downsampling;
s4, marking four different convolution stages in the feature extraction network ResNet 101;
s5, aggregating the output feature map in the step S4 with multi-scale context information through spatial pyramid pooling;
s6, performing double up-sampling on the output characteristic diagram of Conv2_ x through bilinear interpolation, and performing characteristic fusion on the output characteristic diagram of Conv3_ x through a channel attention mechanism;
s7, performing double up-sampling on the output characteristic diagram of the step S6 through bilinear interpolation, and performing characteristic fusion on the output characteristic diagram of the Conv4_ x through a channel attention mechanism to serve as an output characteristic diagram of a channel attention mechanism characteristic fusion module;
and S8, performing splicing operation on the output feature maps of the step S7 and the step S5 in channel dimensions, generating dense feature maps through 3 x 3 convolution, performing 4-time upsampling on the diagnostic map by using bilinear interpolation, reducing the upsampled feature maps to the resolution of the input image, and generating a final prediction result.
Preferably, in step S4, the four different convolution stages are respectively denoted as Conv2_ x, Conv3_ x, Conv4_ x and Conv5_ x, the number of residual structures in Conv2_ x, Conv3_ x, Conv4_ x and Conv5_ x is adjusted to {8,8,9 and 8}, and the adjusted feature extraction network shallow feature map contains more high-level semantic information, so that fusion between the shallow features and the deep features can be effectively guided.
Preferably, in step S5, the spatial pyramid pooling includes four feature extraction paths with different sampling rates and a global average pooling channel, the global average pooling has a global receptive field, the output feature maps of different sampling paths are subjected to channel dimension splicing, and then, 4 times of upsampling is performed by bilinear interpolation to serve as the encoder output.
Therefore, the attention mechanism feature fusion segmentation method for the image has the following beneficial effects:
the method adopts the convolutional neural network to automatically extract the image characteristics without manually selecting the characteristics, so that the image semantic segmentation process can be trained end to end, and compared with the traditional segmentation algorithm, the method is simpler. Compared with the feature map with only the downsampling step length of 4 in the DeepLabv3+, the method adopts the channel attention feature fusion module to aggregate the multi-scale shallow feature map in a cascading mode, obtains richer context information, efficiently recovers the spatial position information of the pixel in a decoder, and generates a finer segmentation result. The invention adopts ResNet101 to extract image characteristics in an encoder, performs spatial pyramid pooling and multi-scale context information, and then performs level-by-level fusion on information in a deep layer characteristic diagram and a shallow layer characteristic diagram in a decoder through a channel attention characteristic fusion module, aiming at reserving more complete low-level characteristic information and generating a clearer target boundary.
The foregoing description is only an overview of the technical solutions of the present invention, and in order to make the technical means of the present invention more clearly understood, the present invention may be implemented in accordance with the content of the description, and in order to make the above and other objects, features, and advantages of the present invention more clearly understood, the following detailed description is given in conjunction with the preferred embodiments, together with the accompanying drawings.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings of the embodiments will be briefly described below.
FIG. 1 is a process diagram of a channel attention map image semantic segmentation algorithm of the present invention;
FIG. 2 is a diagram of a global channel attention calculation process of the present invention;
FIG. 3 is a partial channel attention calculation process diagram of the present invention;
FIG. 4 is a block diagram of a channel attention mechanism feature fusion module of the present invention.
Detailed Description
Other aspects, features and advantages of the present invention will become apparent from the following detailed description, taken in conjunction with the accompanying drawings, which form a part of this specification, and which illustrate, by way of example, the principles of the invention. In the referenced drawings, the same or similar components in different drawings are denoted by the same reference numerals.
The invention provides an attention mechanism feature fusion segmentation method for an image, which comprises the following steps of:
s1, the input image is uniformly cropped to 513 × 513 resolution, and the input image with the original size smaller than 513 × 513 is cropped after the zero padding operation.
S2, the input image size is reduced from 513 × 513 to 257 × 257 with 7 × 7 convolution with a convolution step size of 2.
S3, performing pooling operation with step size of 2 and pooling kernel size of 3 × 3 on the output feature map of step S2, and this down-sampling reduces the feature map size to 129 × 129.
S4, the four different convolution stages in the feature extraction network ResNet101 are respectively denoted as Conv2_ x, Conv3_ x, Conv4_ x, and Conv5_ x. The number of residual structures in Conv2_ x, Conv3_ x, Conv4_ x and Conv5_ x was adjusted to {8,8,9,8}, respectively. The adjusted feature extraction network shallow feature map contains more high-level semantic information, and fusion between shallow features and deep features can be effectively guided.
And S5, aggregating the multi-scale context information of the output feature map in the step S4 through spatial pyramid pooling, wherein the spatial pyramid pooling comprises four feature extraction paths with different sampling rates and a global average pooling channel, and the global average pooling has a global receptive field. And the output characteristic diagrams of different sampling paths are spliced through channel dimensions, and then are subjected to 4 times of upsampling through bilinear interpolation to serve as encoder output.
And S6, performing double up-sampling on the output characteristic diagram of Conv2_ x through bilinear interpolation, and performing characteristic fusion on the output characteristic diagram of Conv3_ x through a channel attention mechanism.
And S7, performing double up-sampling on the output characteristic diagram of the step S6 through bilinear interpolation, and performing characteristic fusion with the output characteristic diagram of Conv4_ x through a channel attention mechanism. And taking the output feature map of the feature fusion module as a channel attention mechanism.
And S8, performing splicing operation on the output feature maps of the step S7 and the step S5 in channel dimensions, generating dense feature maps through 3 x 3 convolution, performing 4-time upsampling on the diagnostic map by using bilinear interpolation, reducing the upsampled feature maps to the resolution of the input image, and generating a final prediction result.
FIG. 1 is a process diagram of the image semantic segmentation algorithm of the channel attention mechanism of the present invention, wherein the image semantic segmentation of the channel attention mechanism mainly comprises the following steps:
step 1, in the stage of an encoder, ResNet101 is used as a backbone network for extracting image features, feature graphs with down-sampling step sizes of 4, 8 and 16 are selected for feature fusion, and low-level information of the image is fully utilized. And the number of residual error structures at different convolution stages in ResNet101 is adjusted, and the proportion of detail information and semantic information in the characteristic diagram is balanced.
And 2, reducing the channel number of the feature map to 256 by convolution operation of 1 multiplied by 1 before feature fusion in a decoder. And performing channel dimension splicing operation on the shallow layer feature map subjected to feature fusion by a channel attention mechanism and an output feature map of an encoder, and enabling a decoder to fully learn the mapping relation between the features and the segmentation target by using 3 x 3 convolution. And finally, performing four-time up-sampling on the feature map, converting the feature map to the resolution of the input image, and outputting the prediction map.
Fig. 2 is a global channel attention calculation process diagram of the present invention, in which a global channel attention module obtains a weight vector with a global receptive field through global average pooling, so that a neural network models the importance of each channel of a feature diagram in a training process, and automatically determines which channels have information on useful information and which channels have information on noise. The global channel attention module mainly comprises the following steps:
and 3, firstly, performing global average pooling on the input feature map, changing the feature map with the scale of H multiplied by W multiplied by C into a vector of 1 multiplied by C, wherein each element in the vector corresponds to one channel in the input feature map and has a global receptive field. The result of the global average pooling is:
Figure BDA0003364576980000061
wherein Z represents a diagonalThe result of global average pooling of the feature maps X with the channel number C, XC(i, j) represents the element in the ith row and jth column on the C channel in the input feature map.
Step 4, taking the vector of the global average pooling in the step 3 as an input to generate a global channel attention matrix MGlobal
MGlobal=FGlobal(XC)×W=W2ReLU(β(W1Z))
Wherein beta represents batch normalization, W1And W2Representing the downscaling and upscaling, respectively, of the feature map using a 1 x 1 convolution operation. MGlobalThe method stores a weight coefficient of the interest degree of each channel of the input image by the semantic segmentation network, and the weight coefficient can be learned in the network training process.
FIG. 3 is a diagram of a local channel attention calculation process of the present invention, where local channel attention learns specialized weight information for pixels at different locations on each channel, as opposed to global channel attention, which generates unique weights for each channel. The local channel attention calculation steps are as follows:
and 5, directly reducing the number of channels of the input feature map to the original 1/16 by adopting 1 × 1 convolution. And modeling the correlation among the channels through convolution, and restoring the number of the channels of the feature map to C. Finally, outputting a weight matrix M with the result that elements at different positions on the same channel are subjected to differential characterizationLocal
MLocal(X)=Conv2(ReLU(β(Conv1(X))))
Conv1And Conv2The 1 × 1 convolution operations of ascending and descending dimensions are respectively performed on the input feature map, and β is batch normalization. The local channel attention changes the pooling window size of the global channel attention, and the pooling operation with the window size of 1 × 1 is adopted in the figure, so that pooling cores with different pooling window sizes can be selected according to specific situations.
FIG. 4 is a block diagram of a channel attention mechanism feature fusion module of the present invention, with input features of the same size passing through a global channel attention module and local viasThe channel attention module models multi-scale channel correlation information to generate a multi-scale channel attention weight matrix Xout
Figure BDA0003364576980000071
The sizes of the input feature map and the output feature map are all H multiplied by W multiplied by C and respectively marked as XinAnd Xout. The input feature map generates a weight matrix M through a global channel attention module and a local channel attention module which are parallelGlobalAnd MLocal. Wherein M isGlobalIs a weight vector of 1 × 1 × C, and MLocalThe weight matrix is H multiplied by W multiplied by C, and the two cannot be directly added because of different scales. Needs to mix MGlobalThe weight matrix expanded to H × W × C along the channel dimension is subjected to the corresponding element summation operation. And then, performing threshold softening on the weight matrix by adopting a Sigmoid function, and mapping coefficients in the weight matrix into a range of (0-1).
The invention adjusts the number of residual error structures in each convolution stage in a feature extraction network ResNet101 in an encoder, and aims to enable a high-resolution low-level feature diagram to bear more high-level semantic information; and the depth separable convolution is adopted to replace the common convolution, so that the model is as light as possible. The low-level feature maps of three different scales are used in a decoder, and feature fusion is carried out in a cascading mode. In feature fusion, a channel attention module is adopted to model the correlation between channels, enhance the feature expression of an interested target and weaken the influence of useless information in a low-level feature map. The validity of the model was verified by the PASCAL VOC 2012 data set. The result shows that AFF-deep Lab can generate a fine segmentation boundary which is better than the classification capability of deep bv3+ on confusable pixels. The average cross-over ratio (mIoU) of the method reaches 81.08%, and is improved by 0.86% compared with DeepLabv3+, and higher semantic segmentation precision is achieved under the condition of not increasing the computational complexity.
While the foregoing is directed to the preferred embodiment of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow.

Claims (3)

1. An attention mechanism feature fusion segmentation method for an image is characterized by comprising the following steps:
s1, uniformly cutting the input image into 513 × 513 resolution, and cutting the input image with the original size smaller than 513 × 513 after zero filling operation;
s2, reducing the input image size from 513 × 513 to 257 × 257 by 7 × 7 convolution with convolution step size 2;
s3, performing pooling operation with step size of 2 and pooling kernel size of 3 × 3 on the output feature map of step S2, and reducing the feature map size to 129 × 129 by downsampling;
s4, marking four different convolution stages in the feature extraction network ResNet 101;
s5, aggregating the output feature map in the step S4 with multi-scale context information through spatial pyramid pooling;
s6, performing double up-sampling on the output characteristic diagram of Conv2_ x through bilinear interpolation, and performing characteristic fusion on the output characteristic diagram of Conv3_ x through a channel attention mechanism;
s7, performing double up-sampling on the output characteristic diagram of the step S6 through bilinear interpolation, and performing characteristic fusion on the output characteristic diagram of the Conv4_ x through a channel attention mechanism to serve as an output characteristic diagram of a channel attention mechanism characteristic fusion module;
and S8, performing splicing operation on the output feature maps of the step S7 and the step S5 in channel dimensions, generating dense feature maps through 3 x 3 convolution, performing 4-time upsampling on the diagnostic map by using bilinear interpolation, reducing the upsampled feature maps to the resolution of the input image, and generating a final prediction result.
2. The method for attention-based feature fusion segmentation of images as claimed in claim 1, wherein in step S4, the four different convolution stages are respectively denoted as Conv2_ x, Conv3_ x, Conv4_ x and Conv5_ x, the number of residual structures in Conv2_ x, Conv3_ x, Conv4_ x and Conv5_ x is adjusted to {8,8,9,8}, and the adjusted feature extraction network shallow feature map contains more high-level semantic information, which can effectively guide the fusion between the shallow features and the deep features.
3. The method of attention-based feature fusion segmentation of images according to claim 1, wherein in step S5, the spatial pyramid pooling includes four feature extraction paths with different sampling rates and a global average pooling channel, the global average pooling has a global receptive field, the output feature maps of different sampling paths are subjected to channel dimension stitching, and then are subjected to 4-fold upsampling by bilinear interpolation to be output by the encoder.
CN202111398830.0A 2021-11-19 2021-11-19 An Image Attention Mechanism Feature Fusion Segmentation Method Pending CN114048822A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111398830.0A CN114048822A (en) 2021-11-19 2021-11-19 An Image Attention Mechanism Feature Fusion Segmentation Method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111398830.0A CN114048822A (en) 2021-11-19 2021-11-19 An Image Attention Mechanism Feature Fusion Segmentation Method

Publications (1)

Publication Number Publication Date
CN114048822A true CN114048822A (en) 2022-02-15

Family

ID=80210798

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111398830.0A Pending CN114048822A (en) 2021-11-19 2021-11-19 An Image Attention Mechanism Feature Fusion Segmentation Method

Country Status (1)

Country Link
CN (1) CN114048822A (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114445430A (en) * 2022-04-08 2022-05-06 暨南大学 Real-time image semantic segmentation method and system based on lightweight multi-scale feature fusion
CN114565770A (en) * 2022-03-23 2022-05-31 中南大学 Image segmentation method and system based on edge-assisted computation and mask attention
CN114708431A (en) * 2022-03-29 2022-07-05 上海大学 Material image segmentation method based on multi-dimensional feature fusion and drawing attention
CN114772208A (en) * 2022-03-31 2022-07-22 东北大学 Non-contact belt tearing detection system and method based on image segmentation
CN114881155A (en) * 2022-05-13 2022-08-09 重庆邮电大学 Fruit image classification method based on deep transfer learning
CN115294412A (en) * 2022-10-10 2022-11-04 临沂大学 Real-time coal and rock segmentation network generation method based on deep learning
CN115564789A (en) * 2022-12-01 2023-01-03 北京矩视智能科技有限公司 Method and device for segmenting defect region of workpiece by cross-level fusion and storage medium
CN115578562A (en) * 2022-10-14 2023-01-06 山西大学 Disc optic cup segmentation method based on MLP and CNN
CN116245891A (en) * 2022-11-18 2023-06-09 三峡大学 Crop area extraction method based on feature interaction and attention decoding
CN116563252A (en) * 2022-06-01 2023-08-08 西南科技大学 A method for segmentation of early esophageal cancer lesions based on feature fusion of dual-branch attention
CN116630343A (en) * 2023-05-23 2023-08-22 青岛理工大学 Rivetless riveting joint profile image segmentation method based on deep learning
CN117789153A (en) * 2024-02-26 2024-03-29 浙江驿公里智能科技有限公司 Automobile oil tank outer cover positioning system and method based on computer vision
CN118038222A (en) * 2024-01-19 2024-05-14 南京邮电大学 Image fusion model and method based on secondary image decomposition and attention mechanism
CN119048530A (en) * 2024-10-28 2024-11-29 江西师范大学 Polyp image segmentation method and system based on detail restoration network

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018035805A1 (en) * 2016-08-25 2018-03-01 Intel Corporation Coupled multi-task fully convolutional networks using multi-scale contextual information and hierarchical hyper-features for semantic image segmentation
WO2020215236A1 (en) * 2019-04-24 2020-10-29 哈尔滨工业大学(深圳) Image semantic segmentation method and system
CN112287940A (en) * 2020-10-30 2021-01-29 西安工程大学 Semantic segmentation method of attention mechanism based on deep learning
CN113486897A (en) * 2021-07-29 2021-10-08 辽宁工程技术大学 Semantic segmentation method for convolution attention mechanism up-sampling decoding
WO2021203832A1 (en) * 2020-04-10 2021-10-14 杭州睿琪软件有限公司 Method and device for removing handwritten content from text image, and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018035805A1 (en) * 2016-08-25 2018-03-01 Intel Corporation Coupled multi-task fully convolutional networks using multi-scale contextual information and hierarchical hyper-features for semantic image segmentation
WO2020215236A1 (en) * 2019-04-24 2020-10-29 哈尔滨工业大学(深圳) Image semantic segmentation method and system
WO2021203832A1 (en) * 2020-04-10 2021-10-14 杭州睿琪软件有限公司 Method and device for removing handwritten content from text image, and storage medium
CN112287940A (en) * 2020-10-30 2021-01-29 西安工程大学 Semantic segmentation method of attention mechanism based on deep learning
CN113486897A (en) * 2021-07-29 2021-10-08 辽宁工程技术大学 Semantic segmentation method for convolution attention mechanism up-sampling decoding

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
曲长波等: "空洞卷积的多尺度语义分割网络", 《计算机工程与应用》, vol. 55, no. 24, 31 December 2019 (2019-12-31) *
翟鹏博;杨浩;宋婷婷;余亢;马龙祥;黄向生;: "结合注意力机制的双路径语义分割", 中国图象图形学报, no. 08, 12 August 2020 (2020-08-12) *

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114565770A (en) * 2022-03-23 2022-05-31 中南大学 Image segmentation method and system based on edge-assisted computation and mask attention
CN114708431A (en) * 2022-03-29 2022-07-05 上海大学 Material image segmentation method based on multi-dimensional feature fusion and drawing attention
CN114772208B (en) * 2022-03-31 2023-06-23 东北大学 Non-contact belt tearing detection system and method based on image segmentation
CN114772208A (en) * 2022-03-31 2022-07-22 东北大学 Non-contact belt tearing detection system and method based on image segmentation
CN114445430A (en) * 2022-04-08 2022-05-06 暨南大学 Real-time image semantic segmentation method and system based on lightweight multi-scale feature fusion
CN114881155A (en) * 2022-05-13 2022-08-09 重庆邮电大学 Fruit image classification method based on deep transfer learning
CN116563252A (en) * 2022-06-01 2023-08-08 西南科技大学 A method for segmentation of early esophageal cancer lesions based on feature fusion of dual-branch attention
CN115294412A (en) * 2022-10-10 2022-11-04 临沂大学 Real-time coal and rock segmentation network generation method based on deep learning
CN115578562A (en) * 2022-10-14 2023-01-06 山西大学 Disc optic cup segmentation method based on MLP and CNN
CN116245891A (en) * 2022-11-18 2023-06-09 三峡大学 Crop area extraction method based on feature interaction and attention decoding
CN115564789A (en) * 2022-12-01 2023-01-03 北京矩视智能科技有限公司 Method and device for segmenting defect region of workpiece by cross-level fusion and storage medium
CN116630343A (en) * 2023-05-23 2023-08-22 青岛理工大学 Rivetless riveting joint profile image segmentation method based on deep learning
CN118038222A (en) * 2024-01-19 2024-05-14 南京邮电大学 Image fusion model and method based on secondary image decomposition and attention mechanism
CN117789153A (en) * 2024-02-26 2024-03-29 浙江驿公里智能科技有限公司 Automobile oil tank outer cover positioning system and method based on computer vision
CN117789153B (en) * 2024-02-26 2024-05-03 浙江驿公里智能科技有限公司 Automobile oil tank outer cover positioning system and method based on computer vision
CN119048530A (en) * 2024-10-28 2024-11-29 江西师范大学 Polyp image segmentation method and system based on detail restoration network

Similar Documents

Publication Publication Date Title
CN114048822A (en) An Image Attention Mechanism Feature Fusion Segmentation Method
CN109299274B (en) Natural scene text detection method based on full convolution neural network
CN110287849B (en) Lightweight depth network image target detection method suitable for raspberry pi
CN110111366B (en) End-to-end optical flow estimation method based on multistage loss
CN108509978B (en) Multi-class target detection method and model based on CNN (CNN) multi-level feature fusion
CN113657388B (en) Image semantic segmentation method for super-resolution reconstruction of fused image
CN113673590B (en) Rain removal method, system and medium based on multi-scale hourglass densely connected network
CN111242288B (en) A Multi-scale Parallel Deep Neural Network Model Construction Method for Lesion Image Segmentation
CN113902915A (en) Semantic segmentation method and system based on low-illumination complex road scene
CN112488025B (en) Double-temporal remote sensing image semantic change detection method based on multi-modal feature fusion
CN113888547A (en) Unsupervised Domain Adaptive Remote Sensing Road Semantic Segmentation Method Based on GAN Network
CN110738697A (en) Monocular depth estimation method based on deep learning
CN114359297B (en) Multi-resolution semantic segmentation method and device based on attention pyramid
CN111191583A (en) Spatial target recognition system and method based on convolutional neural network
CN113313810B (en) 6D attitude parameter calculation method for transparent object
CN112802039B (en) A panoptic segmentation method based on global edge attention
CN113554032B (en) Remote sensing image segmentation method based on multi-path parallel network of high perception
CN114359292A (en) Medical image segmentation method based on multi-scale and attention
CN109767456A (en) A target tracking method based on SiameseFC framework and PFP neural network
CN112950780A (en) Intelligent network map generation method and system based on remote sensing image
CN114565628B (en) Image segmentation method and system based on boundary perception attention
CN113255459B (en) A lane line detection method based on image sequences
CN112990215B (en) Image denoising method, device, equipment and storage medium
CN112270366A (en) Micro target detection method based on self-adaptive multi-feature fusion
CN113436198A (en) Remote sensing image semantic segmentation method for collaborative image super-resolution reconstruction

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
AD01 Patent right deemed abandoned

Effective date of abandoning: 20241213

AD01 Patent right deemed abandoned