Movatterモバイル変換


[0]ホーム

URL:


CN118840615B - Multimodal ship image classification method based on multi-sequence Mamba - Google Patents

Multimodal ship image classification method based on multi-sequence Mamba
Download PDF

Info

Publication number
CN118840615B
CN118840615BCN202411303032.9ACN202411303032ACN118840615BCN 118840615 BCN118840615 BCN 118840615BCN 202411303032 ACN202411303032 ACN 202411303032ACN 118840615 BCN118840615 BCN 118840615B
Authority
CN
China
Prior art keywords
sequence
module
mamba
image
classification
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202411303032.9A
Other languages
Chinese (zh)
Other versions
CN118840615A (en
Inventor
徐从安
吴俊峰
孙显
周伟
高龙
史骏
宿南
林云
蔡卓燃
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Naval Aeronautical University
Original Assignee
Naval Aeronautical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Naval Aeronautical UniversityfiledCriticalNaval Aeronautical University
Priority to CN202411303032.9ApriorityCriticalpatent/CN118840615B/en
Publication of CN118840615ApublicationCriticalpatent/CN118840615A/en
Application grantedgrantedCritical
Publication of CN118840615BpublicationCriticalpatent/CN118840615B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Classifications

Landscapes

Abstract

Translated fromChinese

本发明公开了一种基于多序列曼巴的多模态船只图像分类方法,将同一船只的自然光图像和红外图像同时输入到多模态分类模型中,得到分类结果;所述多模态分类模型包括序列转换模块、交叉注意力曼巴计算模块、交替遍历曼巴计算模块、光谱空间状态融合模块和分类模块。本发明使用差异较大的自然光图像和红外图像作为模型输入,先通过交叉注意力曼巴的计算初步对两种特征进行融合,再通过交叉遍历曼巴计算将两种特征进一步融合,然后利用光谱空间状态融合对两种特征进行融合和解析,从而获得丰富的图像表征,提高了图像分类的准确率。

The present invention discloses a multimodal ship image classification method based on multi-sequence Mamba, wherein the natural light image and infrared image of the same ship are simultaneously input into the multimodal classification model to obtain a classification result; the multimodal classification model comprises a sequence conversion module, a cross-attention Mamba calculation module, an alternating traversal Mamba calculation module, a spectral space state fusion module and a classification module. The present invention uses natural light images and infrared images with large differences as model inputs, first preliminarily fuses the two features through the calculation of cross-attention Mamba, then further fuses the two features through the cross-traversal Mamba calculation, and then fuses and analyzes the two features using the spectral space state fusion, thereby obtaining a rich image representation and improving the accuracy of image classification.

Description

Multi-mode ship image classification method based on multi-sequence Manba
Technical Field
The invention belongs to the field of data identification, and particularly relates to a ship image identification method.
Background
The classification of ship images is one of the key technologies in the field of remote sensing image processing and analysis, and is focused on automatically classifying and identifying different scenes in the ship images. With the rapid development of remote sensing technology, people can acquire a large amount of high-resolution ship image data, and the data cover various scenes such as cities, forests, farmlands and oceans. The data are classified, and the method has important significance for resource management, environment monitoring, disaster emergency and the like.
Traditional ship classification methods rely mainly on manual feature extraction, which has certain limitations in terms of accuracy and efficiency. In recent years, with the development of deep learning, convolutional Neural Networks (CNNs) exhibit strong automatic feature learning capability in a ship image classification task, and the classification precision and efficiency are remarkably improved. However, the existing method is still insufficient in processing complex ship image feature extraction, and the accuracy is low.
Disclosure of Invention
The invention provides a multi-mode ship image classification method based on multi-sequence Manba, which aims to solve the problems of insufficient extracted features and low accuracy of the existing classification method.
The technical scheme of the invention is as follows:
A multimode ship image classification method based on multi-sequence Manba inputs a natural light image and an infrared image of the same ship into a multimode classification model at the same time to obtain a classification result;
The multi-mode classification model comprises a sequence conversion module, a cross attention Manba calculation module, an alternate traversal Manba calculation module, a spectrum space state fusion module and a classification module;
the sequence conversion module is used for respectively converting the input natural light image and the input infrared image into a corresponding natural light token sequence and an infrared token sequence;
The cross attention manba calculation module obtains two groups of first enhancement features based on a natural light token sequence and an infrared token sequence;
The alternate traversal Manba calculation module obtains two groups of second enhancement features based on the two groups of first enhancement features;
the spectrum space state fusion module obtains fusion characteristics based on two groups of second enhancement characteristics;
and the classification module obtains a classification result based on the fusion characteristic.
As a further improvement of the multi-mode ship image classification method based on multi-sequence manba, the sequence conversion module is an image block embedding module;
the natural light token sequence is calculated by the following steps:;
the infrared token sequence is calculated by the following steps:;
Wherein,Representing the activation function in the sequence conversion module,Representing the convolution operation in the sequence conversion module,In order to input the natural light image,Is an input infrared image.
As a further improvement of the multi-mode ship image classification method based on the multi-sequence Manba, a natural light token sequence is setInfrared token sequence,The processing procedure of the cross attention manba calculation module is as follows:
step A-1, calculating a cross attention score:
;
;
Wherein,Representing the normalization operation,Representing the activation function in the cross-attention manba calculation module,Representing a first linear layer in a cross-attention manba calculation module;
Step A-2, pairing sequenceAndForward SSM calculation was performed:
;
;
Wherein,Representing the normalization operation,Representing a second linear layer in the cross-attention manba calculation module,Representing a positive-sequence structured state space model;
step A-3, pairing sequenceAndRespectively performing reverse order arrangement to obtain a sequenceSum sequenceThen pair the sequencesAndReverse order SSM calculation was performed:
;
;
Wherein,Representing a second linear layer in the cross-attention manba calculation module,Representing an inverse structured state space model;
step A-4, calculating two groups of first enhancement features according to the cross attention score, the positive sequence SSM calculation result and the reverse sequence SSM calculation resultAnd,:
;
As a further improvement of the multi-mode ship image classification method based on the multi-sequence Manba, two groups of first enhancement features output by the cross attention Manba calculation module are set as followsAnd,WhereinThe processing procedure of the alternate traversal Manba calculation module is as follows:
Step B-1, performing traversal sequencing on the two groups of first enhancement features to obtain features:
;
Step B-2, pair of characteristicsPerforming attention score calculation to obtain characteristics:
;
Wherein,Representing the normalization operation,Representing alternating traversal of the activation function in the mannba calculation module,Representing alternating traversal of a first linear layer in the mannba computing module;
step B-2, pair of characteristicsThe feature extraction is respectively carried out through the positive sequence SSM and the reverse sequence SSM to obtain the featureAnd:
;
;
Wherein,Representing alternating traversal of the second linear layer in the mannba calculation module,Representing a positive-order structured state space model,Representing an inverse structured state space model;
Step B-3, according to the characteristicsAndComputing enhanced features,:
;
Step B-4, characterizing the enhancement in a reverse manner of the traversal order in step B-1Splitting into two second enhanced featuresAnd,
As a further improvement of the multi-mode ship image classification method based on the multi-sequence Manba, two groups of second enhancement features output by the alternate traversal Manba calculation module are set as followsThe spectrum space state fusion module comprises the following processing procedures:
;
;
;
Wherein,Representing the normalization operation,Representing the first fully-connected layer in the spectral-spatial state fusion module,Representing a second fully-connected layer in the spectral-spatial state fusion module,Representing the activation function in the spectral-spatial state fusion module,Fusion of features for probability output functions
As a further improvement of the multi-mode ship image classification method based on the multi-sequence Manba, the fusion characteristics obtained by the spectrum space state fusion module are set as followsThe processing procedure of the classification module is as follows:
;
Wherein,Representing the fully connected layer of the classification module,As a probability output function, vectorThe element in the image input at present represents the probability that the ship belongs to each category, wherein the category corresponding to the element with the highest probability is the classification result.
As a further improvement of the multi-modal ship image classification method based on the multi-sequence manba, the training process of the multi-modal classification model is as follows:
Constructing a training set, the training set being expressed asWherein, the method comprises the steps of, wherein,For the samples of the i-th group,Representing natural light images in the i-th set of samples,Representing the infrared image in the i-th set of samples,The height of the image is indicated and,The width of the image is represented and,Representing the number of image channels, the vessels in the natural light image and the infrared image in the same sample are the same vessel,Representing the true class of the ship in the i-th set of samples, Z representing the number of samples;
and inputting the samples in the training set into a multi-mode classification model, solving a cross entropy loss function according to the classification result and the real category, and reversely optimizing the network model by using the cross entropy loss function.
Compared with the prior art, the invention has the following beneficial effects:
according to the invention, natural light images and infrared images with larger differences are used as model input, two features are initially fused through the computation of cross attention Manba, so that the two features learn the features of the other party, then the two features are further fused through the computation of cross traversal Manba, the richer image characterization is extracted, and then the two features are fused and analyzed through the fusion of spectral space states, so that the richer image characterization is obtained, and the accuracy of image classification is improved.
Drawings
FIG. 1 is a schematic diagram of a framework of a ship classification model according to the present invention;
FIG. 2 is a schematic diagram of a cross-attention Manba calculation module;
FIG. 3 is a schematic diagram of an alternate traversal Manba calculation module;
fig. 4 is a schematic diagram of a spectral-spatial state fusion module.
Detailed Description
The technical scheme of the invention is described in detail below with reference to the accompanying drawings. It will be apparent that the described embodiments are only some, but not all, embodiments of the invention.
A multi-mode ship image classification method based on multi-sequence Manba inputs natural light images and infrared images of the same ship into a multi-mode classification model at the same time to obtain classification results.
As shown in fig. 1, the multi-modal classification model includes a sequence conversion module, a cross-attention manba calculation module, an alternate traversal manba calculation module, a spectral space state fusion module, and a classification module.
The sequence conversion module is used for respectively converting the input natural light image and the input infrared image into a corresponding natural light token sequence and an infrared token sequence.
In this embodiment, the sequence conversion module is an image block embedding module;
the natural light token sequence is calculated by the following steps:;
the infrared token sequence is calculated by the following steps:;
Wherein,Representing the activation function in the sequence conversion module,Representing the convolution operation in the sequence conversion module,In order to input the natural light image,Is an input infrared image. The natural light token sequence is expressed asThe infrared token sequence is expressed as,
The cross-attention Manbab computation module obtains two sets of first enhancement features based on the natural light token sequence and the infrared token sequence.
As shown in fig. 2, the processing procedure of the cross-attention manba calculation module is as follows:
step A-1, calculating a cross attention score:
;
;
Wherein,Representing the normalization operation,Representing the activation function in the cross-attention manba calculation module,Representing the first linear layer in the cross-attention manba calculation module.
Step A-2, pairing sequenceAndForward SSM calculation was performed:
;
;
Wherein,Representing the normalization operation,Representing a second linear layer in the cross-attention manba calculation module,Representing a positive-order structured state space model.
Step A-3, pairing sequenceAndRespectively performing reverse order arrangement to obtain a sequenceSum sequenceThen pair the sequencesAndReverse order SSM calculation was performed:
;
;
Wherein,Representing a second linear layer in the cross-attention manba calculation module,Representing an inverse structured state space model.
Step A-4, calculating two groups of first enhancement features according to the cross attention score, the positive sequence SSM calculation result and the reverse sequence SSM calculation resultAnd,:
;
The two sets of first enhancement features of the output may be further expressed as:,, Wherein
The alternating traversal Manba calculation module obtains two sets of second enhancement features based on the two sets of first enhancement features.
As shown in fig. 3, the processing procedure of the alternate traversal mannba calculation module is as follows:
Step B-1, performing traversal sequencing on the two groups of first enhancement features to obtain features:
Step B-2, pair of characteristicsPerforming attention score calculation to obtain characteristics:
;
Wherein,Representing the normalization operation,Representing alternating traversal of the activation function in the mannba calculation module,Representing alternating traversal of the first linear layer in the mannba calculation module.
Step B-2, pair of characteristicsThe feature extraction is respectively carried out through the positive sequence SSM and the reverse sequence SSM to obtain the featureAnd:
;
;
Wherein,Representing alternating traversal of the second linear layer in the mannba calculation module,Representing a positive-order structured state space model,Representing an inverse structured state space model.
Step B-3, according to the characteristicsAndComputing enhanced features,:
Step B-4, characterizing the enhancement in a reverse manner of the traversal order in step B-1Splitting into two second enhanced featuresAnd,
The spectrum space state fusion module obtains fusion characteristics based on the two groups of second enhancement characteristics.
As shown in fig. 4, the processing procedure of the spectrum space state fusion module is as follows:
;
;
;
Wherein,Representing the normalization operation,Representing the first fully-connected layer in the spectral-spatial state fusion module,Representing a second fully-connected layer in the spectral-spatial state fusion module,Representing the activation function in the spectral-spatial state fusion module,Fusion of features for probability output functions
The classification module obtains classification results based on the fusion features:
;
Wherein,Representing the fully connected layer of the classification module,As a probability output function, vectorThe element in the image input at present represents the probability that the ship belongs to each category, wherein the category corresponding to the element with the highest probability is the classification result.
Further, the training process of the multi-mode classification model is as follows:
Constructing a training set, the training set being expressed asWherein, the method comprises the steps of, wherein,For the samples of the i-th group,Representing natural light images in the i-th set of samples,Representing the infrared image in the i-th set of samples,The height of the image is indicated and,The width of the image is represented and,Representing the number of image channels, the vessels in the natural light image and the infrared image in the same sample are the same vessel,Representing the true class of the ship in the i-th set of samples, Z represents the number of samples.
And inputting the samples in the training set into a multi-mode classification model, solving a cross entropy loss function according to the classification result and the real category, and reversely optimizing the network model by adopting an Adam optimizer by utilizing the cross entropy loss function.
After training, natural light images and infrared images of the same ship are input into the model at the same time, and classification results are obtained.
It should be noted that it will be apparent to those skilled in the art that the present invention is not limited to the details of the above-described exemplary embodiments, but may be embodied in other specific forms without departing from the spirit or essential characteristics thereof. The scope of the invention is indicated by the appended claims rather than by the foregoing description.

Claims (5)

Translated fromChinese
1.一种基于多序列曼巴的多模态船只图像分类方法,其特征在于:将同一船只的自然光图像和红外图像同时输入到多模态分类模型中,得到分类结果;1. A multimodal ship image classification method based on multi-sequence Mamba, characterized in that: the natural light image and infrared image of the same ship are simultaneously input into the multimodal classification model to obtain the classification result;所述多模态分类模型包括序列转换模块、交叉注意力曼巴计算模块、交替遍历曼巴计算模块、光谱空间状态融合模块和分类模块;The multimodal classification model includes a sequence conversion module, a cross-attention Mamba calculation module, an alternating traversal Mamba calculation module, a spectral space state fusion module and a classification module;所述序列转换模块用于将输入的自然光图像和红外图像分别转换为对应的自然光token序列和红外token序列;The sequence conversion module is used to convert the input natural light image and infrared image into corresponding natural light token sequence and infrared token sequence respectively;所述交叉注意力曼巴计算模块基于自然光token序列和红外token序列得到两组第一增强特征;The cross-attention Mamba calculation module obtains two sets of first enhanced features based on the natural light token sequence and the infrared token sequence;设自然光token序列,红外token序列,所述交叉注意力曼巴计算模块的处理过程为:Set the natural light token sequence , infrared token sequence , , the processing process of the cross attention Mamba calculation module is:步骤A-1、计算交叉注意力分数Step A-1: Calculate the cross attention score : ; ;其中,代表归一化操作,代表交叉注意力曼巴计算模块中的激活函数,代表交叉注意力曼巴计算模块中的第一线性层;in, represents the normalization operation, represents the activation function in the cross-attention Mamba calculation module, Represents the first linear layer in the cross-attention Mamba computation module;步骤A-2、对序列进行正序SSM计算:Step A-2: Sequence and Perform positive sequence SSM calculation: ; ;其中,代表归一化操作,代表交叉注意力曼巴计算模块中的第二线性层,代表正序结构化状态空间模型;in, represents the normalization operation, represents the second linear layer in the cross-attention Mamba computation module, represents the positive sequence structured state space model;步骤A-3、对序列分别进行倒序排列,得到序列和序列;然后对序列进行逆序SSM计算:Step A-3: Sequence and Arrange them in reverse order to get the sequence and sequence ; Then the sequence and Perform reverse SSM calculation: ; ;其中,代表交叉注意力曼巴计算模块中的第二线性层,代表逆序结构化状态空间模型;in, represents the second linear layer in the cross-attention Mamba computation module, stands for reverse structured state space model;步骤A-4、根据交叉注意力分数、正序SSM计算结果和逆序SSM计算结果计算两组第一增强特征Step A-4: Calculate two sets of first enhanced features based on the cross attention score, the forward SSM calculation result, and the reverse SSM calculation result and , : ; ;所述交替遍历曼巴计算模块基于两组第一增强特征得到两组第二增强特征;The alternating traversal Mamba calculation module obtains two groups of second enhanced features based on the two groups of first enhanced features;设交叉注意力曼巴计算模块输出的两组第一增强特征为,其中,所述交替遍历曼巴计算模块的处理过程为:Assume that the two sets of first enhanced features output by the cross-attention Mamba calculation module are and , ,in , the processing process of alternatingly traversing the Mamba calculation module is:步骤B-1、将两组第一增强特征进行遍历排序,得到特征Step B-1: Traverse and sort the two sets of first enhanced features to obtain features : ;步骤B-2、对特征进行注意力分数计算,得到特征Step B-2: Features Calculate the attention score and get the features : ;其中,代表归一化操作,代表交替遍历曼巴计算模块中的激活函数,代表交替遍历曼巴计算模块中的第一线性层;in, represents the normalization operation, represents the activation function in the alternate traversal Mamba calculation module, Represents alternating traversal of the first linear layer in the Mamba calculation module;步骤B-2、对特征分别通过正序SSM和逆序SSM进行特征提取,得到特征Step B-2: Features Feature extraction is performed through positive order SSM and reverse order SSM respectively to obtain the feature and : ; ;其中,代表交替遍历曼巴计算模块中的第二线性层,代表正序结构化状态空间模型,代表逆序结构化状态空间模型;in, represents the alternating traversal of the second linear layer in the Mamba computation module, represents the positive sequence structured state space model, stands for reverse structured state space model;步骤B-3、根据特征计算增强特征Step B-3: Based on characteristics , and Compute Enhanced Features , : ;步骤B-4、按步骤B-1中遍历排序的逆向方式将增强特征拆分成两个第二增强特征Step B-4: Enhance the features by traversing and sorting in the reverse way in step B-1 Split into two second enhancement features and , ;所述光谱空间状态融合模块基于两组第二增强特征得到融合特征;The spectrum-space state fusion module obtains fusion features based on two sets of second enhanced features;所述分类模块基于所述融合特征得到分类结果。The classification module obtains a classification result based on the fusion features.2.如权利要求1所述的基于多序列曼巴的多模态船只图像分类方法,其特征在于:所述序列转换模块为图像块嵌入模块;2. The multimodal ship image classification method based on multi-sequence Mamba as claimed in claim 1, characterized in that: the sequence conversion module is an image block embedding module;自然光token序列的计算方式为:The calculation method of natural light token sequence is: ;红外token序列的计算方式为:The calculation method of infrared token sequence is: ;其中,代表序列转换模块中的激活函数,代表序列转换模块中的卷积操作,为输入的自然光图像,为输入的红外图像。in, represents the activation function in the sequence conversion module, represents the convolution operation in the sequence conversion module, is the input natural light image, is the input infrared image.3.如权利要求1所述的基于多序列曼巴的多模态船只图像分类方法,其特征在于:设交替遍历曼巴计算模块输出的两组第二增强特征为,光谱空间状态融合模块的处理过程为:3. The multimodal ship image classification method based on multi-sequence Mamba as claimed in claim 1, characterized in that: the two sets of second enhanced features output by the alternately traversed Mamba calculation module are set as , the processing process of the spectrum space state fusion module is: ; ; ;其中,代表归一化操作,代表光谱空间状态融合模块中的第一全连接层,代表光谱空间状态融合模块中的第二全连接层,代表光谱空间状态融合模块中的激活函数,为概率输出函数,融合特征in, represents the normalization operation, represents the first fully connected layer in the spectral-spatial state fusion module, represents the second fully connected layer in the spectral-spatial state fusion module, represents the activation function in the spectral-spatial state fusion module, is the probability output function, fusion feature .4.如权利要求1所述的基于多序列曼巴的多模态船只图像分类方法,其特征在于:设光谱空间状态融合模块得到的融合特征为,则分类模块的处理过程为:4. The multimodal ship image classification method based on multi-sequence Mamba as claimed in claim 1 is characterized in that: the fusion feature obtained by the spectrum-space state fusion module is , then the processing process of the classification module is: ;其中,代表分类模块的全连接层,为概率输出函数,向量中的各元素表示当前输入的图像中的船只属于各类别的概率,其中概率最大的元素对应的类别即为分类结果。in, represents the fully connected layer of the classification module, is the probability output function, vector Each element in represents the probability that the ship in the current input image belongs to each category, and the category corresponding to the element with the largest probability is the classification result.5.如权利要求1至4任一所述的基于多序列曼巴的多模态船只图像分类方法,其特征在于,所述多模态分类模型的训练过程为:5. The multimodal ship image classification method based on multi-sequence Mamba according to any one of claims 1 to 4, characterized in that the training process of the multimodal classification model is:构建训练集,所述训练集表示为,其中,为第i组样本,表示第i组样本中的自然光图像,表示第i组样本中的红外图像,表示图像高度,表示图像宽度,表示图像通道数,同一样本中的自然光图像和红外图像中的船只为同一船只,表示第i组样本中的船只的真实类别,Z表示样本数量;Construct a training set, which is represented by ,in, is thei- th group of samples, represents the natural light image in thei- th group of samples, represents the infrared image in the i- th group of samples, Indicates the image height, Indicates the image width, Indicates the number of image channels. The ships in the natural light image and infrared image of the same sample are the same ship. represents the true category of the ship in thei- th group of samples,and Z represents the number of samples;将训练集中的样本输入多模态分类模型,根据分类结果和真实类别求出交叉熵损失函数,再利用交叉熵损失函数反向优化网络模型。The samples in the training set are input into the multimodal classification model, the cross entropy loss function is calculated according to the classification results and the true categories, and then the cross entropy loss function is used to reversely optimize the network model.
CN202411303032.9A2024-09-192024-09-19 Multimodal ship image classification method based on multi-sequence MambaActiveCN118840615B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202411303032.9ACN118840615B (en)2024-09-192024-09-19 Multimodal ship image classification method based on multi-sequence Mamba

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202411303032.9ACN118840615B (en)2024-09-192024-09-19 Multimodal ship image classification method based on multi-sequence Mamba

Publications (2)

Publication NumberPublication Date
CN118840615A CN118840615A (en)2024-10-25
CN118840615Btrue CN118840615B (en)2025-01-21

Family

ID=93139441

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202411303032.9AActiveCN118840615B (en)2024-09-192024-09-19 Multimodal ship image classification method based on multi-sequence Mamba

Country Status (1)

CountryLink
CN (1)CN118840615B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN119580920A (en)*2024-11-142025-03-07中南林业科技大学 A medical image description method based on state space model

Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN115082434A (en)*2022-07-212022-09-20浙江华是科技股份有限公司Multi-source feature-based magnetic core defect detection model training method and system
CN116311387A (en)*2023-05-252023-06-23浙江工业大学 A cross-modal person re-identification method based on feature intersection

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
EP2808760B1 (en)*2013-05-292023-08-16Dassault SystèmesBody posture tracking
CN115240121B (en)*2022-09-222023-01-03之江实验室Joint modeling method and device for enhancing local features of pedestrians
CN115272776B (en)*2022-09-262023-01-20山东锋士信息技术有限公司Hyperspectral image classification method based on double-path convolution and double attention and storage medium
CN117115876A (en)*2023-04-132023-11-24重庆交通大学Face automatic identification method and system adapting to visible light and near infrared
CN117372763B (en)*2023-10-162024-11-29中国人民解放军海军航空大学Remote sensing scene classification method based on mask attention mechanism
CN118447244A (en)*2024-04-292024-08-06南通大学 Brain tumor image segmentation method based on multi-scale convolution and Mamba structure
CN118537369A (en)*2024-05-152024-08-23中国石油大学(华东)Single-target tracking network based on Vmamba and transducer combination
CN118470641B (en)*2024-05-222024-10-22鸣飞伟业技术有限公司 Ship overload determination method and device based on image recognition
CN118295029B (en)*2024-06-032024-08-06中国石油大学(华东)Seismic data denoising method integrating self-attention and Mamba architecture

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN115082434A (en)*2022-07-212022-09-20浙江华是科技股份有限公司Multi-source feature-based magnetic core defect detection model training method and system
CN116311387A (en)*2023-05-252023-06-23浙江工业大学 A cross-modal person re-identification method based on feature intersection

Also Published As

Publication numberPublication date
CN118840615A (en)2024-10-25

Similar Documents

PublicationPublication DateTitle
CN113255755A (en)Multi-modal emotion classification method based on heterogeneous fusion network
CN116383517B (en) Multimodal rumor detection method and system with enhanced dynamic propagation features
CN115631365A (en) A Cross-Modal Contrastive Zero-Shot Learning Method Fused with Knowledge Graph
WO2021243903A1 (en)Method and system for transforming natural language into structured query language
CN108229588B (en)Machine learning identification method based on deep learning
CN118840615B (en) Multimodal ship image classification method based on multi-sequence Mamba
CN114170411B (en) A method for image emotion recognition by integrating multi-scale information
CN116310305B (en) A semantic segmentation model with encoder-decoder structure based on tensor and second-order covariance attention mechanism
CN117516937A (en) Unknown fault detection method of rolling bearing based on multi-modal feature fusion enhancement
CN117557775A (en)Substation power equipment detection method and system based on infrared and visible light fusion
CN115908793A (en) A Semantic Segmentation Model for Encoding and Decoding Structures Based on Positional Attention Mechanism
CN117633674A (en)Multi-mode emotion analysis method based on causal gating attention mechanism
CN112163490A (en)Target detection method based on scene picture
CN114612666A (en) A RGB-D Semantic Segmentation Method Based on Multimodal Contrastive Learning
CN110414338B (en) Pedestrian re-identification method based on sparse attention network
CN113343760A (en)Human behavior recognition method based on multi-scale characteristic neural network
CN117688509A (en)Multi-mode false news detection method based on multi-level fusion and attention mechanism
CN112365139A (en)Crowd danger degree analysis method under graph convolution neural network
CN112990196A (en)Scene character recognition method and system based on hyper-parameter search and two-stage training
CN118114105A (en) A multimodal emotion recognition method and system based on contrastive learning and Transformer structure
CN116206327A (en)Image classification method based on online knowledge distillation
CN119445544A (en) A rock image classification method based on attention guidance and multi-feature fusion
CN119068502A (en) A method for authenticity identification of multimodal information in social media
CN109766918A (en) Salient object detection method based on multi-level context information fusion
CN115512357A (en) A zero-sample Chinese character recognition method based on component splitting

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp