Movatterモバイル変換


[0]ホーム

URL:


CN112308211B - Domain increment method based on meta learning - Google Patents

Domain increment method based on meta learning
Download PDF

Info

Publication number
CN112308211B
CN112308211BCN202011186818.9ACN202011186818ACN112308211BCN 112308211 BCN112308211 BCN 112308211BCN 202011186818 ACN202011186818 ACN 202011186818ACN 112308211 BCN112308211 BCN 112308211B
Authority
CN
China
Prior art keywords
model
training
old
image data
new
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011186818.9A
Other languages
Chinese (zh)
Other versions
CN112308211A (en
Inventor
王杰龙
安竹林
程坦
徐勇军
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhongke Xiamen Data Intelligence Research Institute
Original Assignee
Zhongke Xiamen Data Intelligence Research Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhongke Xiamen Data Intelligence Research InstitutefiledCriticalZhongke Xiamen Data Intelligence Research Institute
Priority to CN202011186818.9ApriorityCriticalpatent/CN112308211B/en
Publication of CN112308211ApublicationCriticalpatent/CN112308211A/en
Application grantedgrantedCritical
Publication of CN112308211BpublicationCriticalpatent/CN112308211B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Classifications

Landscapes

Abstract

The invention discloses a field increment method based on meta learning, which comprises the following steps: s1, constructing a pre-training model; s2, training an old model by using a pre-training model; s3, training a new model. The invention adopts 5% of memory image data and new image data which are randomly reserved to mix and finely tune to train a new model, and simultaneously adopts a cross entropy loss function and a knowledge distillation loss function to jointly guide the study of the new model, so that the classification knowledge of the data in the new field is learned while the classification knowledge in the old field is remembered, and the expenditure of data storage and training time is greatly reduced.

Description

Domain increment method based on meta learning
Technical Field
The invention relates to the technical field of computers, in particular to a field increment method based on meta learning.
Background
With the rise of deep learning, the object classification method based on the convolutional neural network is rapidly developed, and the recognition accuracy is greatly improved. However, convolutional neural network-based approaches also have drawbacks: when the tested picture data distribution is inconsistent with the training picture data distribution, such as illumination, background, gesture and the like, the accuracy of the model is reduced. Therefore, when new domain data, i.e., data inconsistent with the original training data distribution, appears, it is necessary that the model be able to learn new domain classifications incrementally, i.e., learn new domain knowledge classifications while remembering classifications of old domain data.
At present, the most intuitive field increment learning method is to continue training a model by using data in a new field, but the method often has the condition that the precision cannot meet the requirement: if the training is insufficient, the data accuracy rate for the new field is not high; if training is excessive, the accuracy of the data for the old domain is reduced, and the data are difficult to reconcile. If the old domain data and the new domain data are directly mixed and retrained into the convolutional neural network, the cost of data storage and training time is huge, and particularly, the cost is larger and larger as the new domain data are larger and larger in practice. Therefore, it is particularly important to find a domain increment recognition method that can obtain high-precision performance at low overhead cost.
Disclosure of Invention
In order to solve the problems, the invention provides a field increment method based on meta learning.
The invention adopts the following technical scheme:
a domain increment method based on meta learning comprises the following steps:
s1, constructing a pre-training model: selecting a plurality of public data sets as metadata by using a meta learning method iTAML, constructing a meta task and learning a pre-training model to obtain parameters phi of the pre-training model, wherein the pre-training model is a convolutional neural classification network;
s2, training an old model by using the pre-training model: constructing a classification model of the same type as the pre-training model as an old model, importing parameters phi of the pre-training model into the old model, and guiding old image data D by using a cross entropy loss functionold Training the old model, and randomly sampling and retaining 5% of old image data D after training is finishedold As the memory image data Dmemory
S3, training a new model: using the memory image data Dmemory And new image data Dnew Training the old model together in a hybrid manner for new image data Dnew Guiding model learning using cross entropy loss function for memorizing image data Dmemory Model learning is guided jointly by using the cross entropy loss function and the knowledge distillation loss function, thereby obtainingTo the new model.
Further, the convolutional neural classification network is one of VGG, resNet, mobileNet, denseNet or SENet.
Further, in step S1, the training process of the meta learning method iTAML is incremental, T stages are trained altogether, T is the total task number, and T represents the T-th task;
when t=1, the data of task 1 is normally trained by using a cross entropy loss formula, and a pre-training model parameter phi is obtained1 The cross entropy loss formula is as follows:
wherein D ist Representing a data set belonging to a t-th task, a total of N samples, xi One of them, pi Representation model pair xi Predicted value of yi Representing a true tag value;
when t is more than or equal to 2, initializing the parameter to be the trained parameter phi in the previous stagebase =φt-1 Respectively taking out task 1, task 2, … and task t, and taking out t task data to obtain phibase Updating and optimizing the initial parameters by using cross entropy loss to obtain temporary parameters phi of the corresponding tasks12 ,…φt Then update phibase Obtaining the final result parameter phi at this stage when the loss does not dropt =φbase The update phibase The following formula is adopted:
finally, to obtain phiT As parameters of the pre-training model.
Further, in step S3, the model learning is guided by combining the cross entropy loss function and the knowledge distillation loss function, and the formula is as follows: loss=loss_ce+loss_distil, where loss_ce represents cross entropy Loss, loss_distil represents knowledge distillation Loss,
and the solution formula for loss_ce is as follows:
wherein x isi ∈Dmemory ∪Dnew Samples representing memory image data or new image data, N in total, pi Representation model pair xi Predicted value of yi Representing a true tag value;
and the solution formula of loss_distill is as follows:
wherein x isi ∈Dmemory A total of N, q samples representing the memory image datai Is the old model about data xi Predicted value of pi Is a model pair x in trainingi Is a predicted value of (a).
After the technical scheme is adopted, compared with the background technology, the invention has the following advantages:
1. unlike conventional pre-training models using large public data sets, meta-learning pre-training models have the advantage of fast learning new task data, and can train target task data in less time;
2. metadata of the pre-training model does not need to be stored, so that training time is not increased along with the increase of new image data;
3. the new model is trained by adopting 5% of memory image data and new image data which are randomly reserved in a mixed fine tuning mode, and simultaneously, the learning of the new model is guided by adopting a cross entropy loss function and a knowledge distillation loss function in a combined mode, so that the classification knowledge of the data in the new field is learned while the classification knowledge in the old field is remembered, the expenditure of data storage and training time is greatly reduced, and meanwhile, the accuracy of the model after the new field increment is introduced is ensured.
Drawings
FIG. 1 is a schematic flow chart of the method of the present invention.
Detailed Description
The present invention will be described in further detail with reference to the drawings and examples, in order to make the objects, technical solutions and advantages of the present invention more apparent. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
Examples
In the present embodiment, a batch of old image data D is givenold ,Dold Consists of Mobye and gold wool and a new image data Dnew ,Dnew The method consists of a small Huang Che and a Husky, and the aim of the embodiment is to achieve higher accuracy for both new and old image data. As shown in fig. 1, the method of this embodiment is as follows:
s1, constructing a pre-training model: the meta-learning method iTAML is utilized, a plurality of disclosed data sets are selected as metadata, meta-tasks are constructed, a pre-training model is learned, such as selecting an airplane and a bird in the cifar10 data set (task 1), driving a truck and a deer (task 2), and an automobile and a horse (task 3), the classification model structure selects mobilenet v2, and parameters phi of the pre-training model are obtained, the pre-training model is a convolutional neural classification network, and it is pointed out that iTAML is independent of the model, and any convolutional neural classification network such as VGG, resNet, mobileNet, denseNet or SENet can be selected.
In step S1, the training process of the meta learning method iTAML is incremental, T stages are trained altogether, T is the total task number, T represents the T-th task, and in this embodiment, 3 tasks are taken as an example:
when t=1, the data of task 1 is normally trained by using a cross entropy loss formula, and a pre-training model parameter phi is obtained1 The cross entropy loss formula is as follows:
wherein D ist Representing a data set belonging to a t-th task, a total of N samples, xi One of them, pi Representation model pair xi Predicted value of yi Representing a true tag value;
when t is more than or equal to 2, initializing the parameter to be the trained parameter phi in the previous stagebase =φt-1 Respectively taking out task 1, task 2, … and task t, and taking out t task data to obtain phibase Updating and optimizing the initial parameters by using cross entropy loss to obtain temporary parameters phi of the corresponding tasks12 ,…φt Then update phibase Obtaining the final result parameter phi at this stage when the loss does not dropt =φbase The update phibase The following formula is adopted:
finally, to obtain phi3 As parameters of the pre-training model.
S2, training an old model by using the pre-training model: constructing a classification model of the same type as the pre-training model as an old model, importing parameters phi of a feature extraction layer in the pre-training model into the old model, and guiding old image data D by using a cross entropy loss functionold Training the old model, and randomly sampling and retaining 5% of old image data D after training is finishedold As the memory image data Dmemory
Through this step, the old model is applied to the old image data Dold The distributed data has better classification accuracy, but for unknown data which is not distributed in accordance with the data, the accuracy cannot be ensured.
S3, training a new model: using the memory image data Dmemory And new image data Dnew Training the old model together in a hybrid manner for new image data Dnew Guiding model learning using cross entropy loss function for memorizing image data Dmemory And using the cross entropy loss function and the knowledge distillation loss function to jointly guide model learning, thereby obtaining a new model.
In the step S3, the integral loss is adopted when the model learning is guided by combining the cross entropy loss function and the knowledge distillation loss function, and the formula is as follows: loss=loss_ce+loss_distil, where loss_ce represents cross entropy Loss, loss_distil represents knowledge distillation Loss,
and the solution formula for loss_ce is as follows:
wherein x isi ∈Dmemory ∪Dnew Samples representing memory image data or new image data, N in total, pi Representation model pair xi Predicted value of yi Representing a true tag value;
and the solution formula of loss_distill is as follows:
wherein x isi ∈Dmemory A total of N, q samples representing the memory image datai Is the old model about data xi Predicted value of pi Is a model pair x in trainingi Is a predicted value of (a).
The new model is trained by adopting 5% of memory image data and new image data which are randomly reserved in a mixed fine tuning mode, and simultaneously, the learning of the new model is guided by adopting a cross entropy loss function and a knowledge distillation loss function in a combined mode, so that the classification knowledge of the old field is remembered, the classification knowledge of the new field data is learned, the expenditure of data storage and training time is greatly reduced, and meanwhile, the accuracy of the model after the new field increment is introduced is ensured.
The present invention is not limited to the above-mentioned embodiments, and any changes or substitutions that can be easily understood by those skilled in the art within the technical scope of the present invention are intended to be included in the scope of the present invention. Therefore, the protection scope of the present invention should be subject to the protection scope of the claims.

Claims (2)

CN202011186818.9A2020-10-292020-10-29Domain increment method based on meta learningActiveCN112308211B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202011186818.9ACN112308211B (en)2020-10-292020-10-29Domain increment method based on meta learning

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202011186818.9ACN112308211B (en)2020-10-292020-10-29Domain increment method based on meta learning

Publications (2)

Publication NumberPublication Date
CN112308211A CN112308211A (en)2021-02-02
CN112308211Btrue CN112308211B (en)2024-03-08

Family

ID=74332295

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202011186818.9AActiveCN112308211B (en)2020-10-292020-10-29Domain increment method based on meta learning

Country Status (1)

CountryLink
CN (1)CN112308211B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN112990280B (en)*2021-03-012023-08-25华南理工大学Class increment classification method, system, device and medium for image big data
CN113128419B (en)*2021-04-232023-12-05京东鲲鹏(江苏)科技有限公司Obstacle recognition method and device, electronic equipment and storage medium
CN113139487A (en)*2021-04-292021-07-20中国地质大学(武汉)Building feature learning method oriented to remote sensing image incremental data
CN113222020B (en)*2021-05-132024-05-24中科(厦门)数据智能研究院Domain increment learning method based on data conversion and knowledge distillation
CN113486185B (en)*2021-09-072021-11-23中建电子商务有限责任公司Knowledge distillation method based on joint training, processor and storage medium
CN114297491B (en)*2021-12-282024-10-25北京三快在线科技有限公司Model training method, information recommending method and device

Citations (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
GB201908574D0 (en)*2019-06-142019-07-31Vision Semantics LtdOptimised machine learning
CN111199242A (en)*2019-12-182020-05-26浙江工业大学Image increment learning method based on dynamic correction vector
CN111563663A (en)*2020-04-162020-08-21五邑大学Robot, service quality evaluation method and system
CN111681059A (en)*2020-08-142020-09-18支付宝(杭州)信息技术有限公司 Training method and device for behavior prediction model

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN110008842A (en)*2019-03-092019-07-12同济大学A kind of pedestrian's recognition methods again for more losing Fusion Model based on depth

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
GB201908574D0 (en)*2019-06-142019-07-31Vision Semantics LtdOptimised machine learning
CN111199242A (en)*2019-12-182020-05-26浙江工业大学Image increment learning method based on dynamic correction vector
CN111563663A (en)*2020-04-162020-08-21五邑大学Robot, service quality evaluation method and system
CN111681059A (en)*2020-08-142020-09-18支付宝(杭州)信息技术有限公司 Training method and device for behavior prediction model

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Jathushan Rajasegaran ; Salman Khan ; .iTAML: An Incremental Task-Agnostic Meta-learning Approach.《2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)》.2020,全文.*
基于实例的领域适应增量学习方法研究;潘振春;《中国优秀硕士学位论文全文数据库 信息科技辑》;20170715;全文*
基于少量样本的深度学习图像修复算法研究;徐勇军;《中国优秀硕士学位论文全文数据库 信息科技辑》;20200215;全文*

Also Published As

Publication numberPublication date
CN112308211A (en)2021-02-02

Similar Documents

PublicationPublication DateTitle
CN112308211B (en)Domain increment method based on meta learning
CN111368874B (en) An Incremental Learning Method of Image Category Based on Single Classification Technology
US11335328B2 (en)Unsupervised learning of semantic audio representations
CN114708434B (en)Cross-domain remote sensing image semantic segmentation method based on iterative intra-domain adaptation and self-training
CN111210000B (en)Modulation signal increment learning method based on fixed features
CN114332621B (en)Disease and pest identification method and system based on multi-model feature fusion
CN110728294A (en)Cross-domain image classification model construction method and device based on transfer learning
CN111241992B (en)Face recognition model construction method, recognition method, device, equipment and storage medium
CN113469186B (en) A cross-domain transfer image segmentation method based on a small number of point annotations
WO2020232840A1 (en)Vehicle multi-attribute identification method and device employing neural network structure search, and medium
CN111753995B (en)Local interpretable method based on gradient lifting tree
CN112330052A (en)Distribution transformer load prediction method
US20220215679A1 (en)Method of determining a density of cells in a cell image, electronic device, and storage medium
CN114781611A (en)Natural language processing method, language model training method and related equipment
CN115359315A (en)Incremental small sample SAR target identification method based on angular feature separation and retention
CN114022754B (en) A few-shot image recognition method combined with contrastive learning
CN118569260B (en) Small-sample named entity recognition method for joint operation of reservoir groups
CN116798081B (en) A small sample radio frequency fingerprint recognition method
CN117272999A (en)Model training method and device based on class incremental learning, equipment and storage medium
CN114495175B (en)Pedestrian re-identification model creation method, device, equipment and storage medium
CN112163617B (en) Unlabeled numerical type feature classification method, device, equipment and readable storage medium
CN113537290B (en) An image matching method based on ultra-high-dimensional data element clustering
CN116796842A (en)Model reasoning method based on incremental learning and electronic equipment
CN115292535A (en)Hierarchical vehicle image retrieval method based on depth polarization Hash
CN112990430A (en)Group division method and system based on long-time and short-time memory network

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
CB02Change of applicant information

Country or region after:China

Address after:No. 208-3, Fengqi Road, phase III, software park, Xiamen City, Fujian Province, 361000

Applicant after:Zhongke (Xiamen) data Intelligence Research Institute

Address before:No. 208-3, Fengqi Road, phase III, software park, Xiamen City, Fujian Province, 361000

Applicant before:Xiamen Institute of data intelligence, Institute of computing technology, Chinese Academy of Sciences

Country or region before:China

CB02Change of applicant information
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp